feat: Initial commit
This commit is contained in:
62
.github/workflows/build.yaml
vendored
Normal file
62
.github/workflows/build.yaml
vendored
Normal file
@@ -0,0 +1,62 @@
|
||||
name: build
|
||||
|
||||
on:
|
||||
push:
|
||||
branches: [ master ]
|
||||
pull_request:
|
||||
branches: [ master ]
|
||||
|
||||
env:
|
||||
CARGO_TERM_COLOR: always
|
||||
|
||||
jobs:
|
||||
checks-matrix:
|
||||
runs-on: ubuntu-latest
|
||||
outputs:
|
||||
matrix: ${{ steps.set-matrix.outputs.matrix }}
|
||||
steps:
|
||||
- uses: actions/checkout@v4
|
||||
- uses: DeterminateSystems/nix-installer-action@main
|
||||
- uses: DeterminateSystems/magic-nix-cache-action@main
|
||||
- id: set-matrix
|
||||
name: Generate Nix Matrix
|
||||
run: |
|
||||
set -Eeu
|
||||
matrix="$(nix eval --json '.#githubActions.matrix')"
|
||||
echo "matrix=$matrix" >> "$GITHUB_OUTPUT"
|
||||
|
||||
checks-build:
|
||||
needs: checks-matrix
|
||||
runs-on: ${{ matrix.os }}
|
||||
strategy:
|
||||
matrix: ${{fromJSON(needs.checks-matrix.outputs.matrix)}}
|
||||
steps:
|
||||
- uses: actions/checkout@v4
|
||||
- uses: DeterminateSystems/nix-installer-action@main
|
||||
- uses: DeterminateSystems/magic-nix-cache-action@main
|
||||
- run: nix build -L '.#${{ matrix.attr }}'
|
||||
|
||||
codecov:
|
||||
runs-on: ubuntu-latest
|
||||
permissions:
|
||||
id-token: "write"
|
||||
contents: "read"
|
||||
|
||||
steps:
|
||||
- uses: actions/checkout@v4
|
||||
- uses: DeterminateSystems/nix-installer-action@main
|
||||
- uses: DeterminateSystems/magic-nix-cache-action@main
|
||||
|
||||
- name: Run codecov
|
||||
run: nix build .#checks.x86_64-linux.hello-llvm-cov
|
||||
|
||||
- name: Upload coverage reports to Codecov
|
||||
uses: codecov/codecov-action@v4.0.1
|
||||
with:
|
||||
flags: unittests
|
||||
name: codecov-hello
|
||||
fail_ci_if_error: true
|
||||
token: ${{ secrets.CODECOV_TOKEN }}
|
||||
files: ./result
|
||||
verbose: true
|
||||
|
||||
38
.github/workflows/docs.yaml
vendored
Normal file
38
.github/workflows/docs.yaml
vendored
Normal file
@@ -0,0 +1,38 @@
|
||||
name: docs
|
||||
|
||||
on:
|
||||
push:
|
||||
branches: [ master ]
|
||||
|
||||
env:
|
||||
CARGO_TERM_COLOR: always
|
||||
|
||||
jobs:
|
||||
docs:
|
||||
runs-on: ubuntu-latest
|
||||
permissions:
|
||||
id-token: "write"
|
||||
contents: "read"
|
||||
pages: "write"
|
||||
|
||||
steps:
|
||||
- uses: actions/checkout@v4
|
||||
- uses: DeterminateSystems/nix-installer-action@main
|
||||
- uses: DeterminateSystems/magic-nix-cache-action@main
|
||||
- uses: DeterminateSystems/flake-checker-action@main
|
||||
|
||||
- name: Generate docs
|
||||
run: nix build .#checks.x86_64-linux.hello-docs
|
||||
|
||||
- name: Setup Pages
|
||||
uses: actions/configure-pages@v5
|
||||
|
||||
- name: Upload artifact
|
||||
uses: actions/upload-pages-artifact@v3
|
||||
with:
|
||||
path: result/share/doc
|
||||
|
||||
- name: Deploy to gh-pages
|
||||
id: deployment
|
||||
uses: actions/deploy-pages@v4
|
||||
|
||||
4
.gitignore
vendored
Normal file
4
.gitignore
vendored
Normal file
@@ -0,0 +1,4 @@
|
||||
/result
|
||||
/target
|
||||
.direnv
|
||||
*.jpg
|
||||
2559
Cargo.lock
generated
Normal file
2559
Cargo.lock
generated
Normal file
File diff suppressed because it is too large
Load Diff
44
Cargo.toml
Normal file
44
Cargo.toml
Normal file
@@ -0,0 +1,44 @@
|
||||
[package]
|
||||
name = "detector"
|
||||
version = "0.1.0"
|
||||
edition = "2024"
|
||||
license = "MIT"
|
||||
|
||||
# See more keys and their definitions at https://doc.rust-lang.org/cargo/reference/manifest.html
|
||||
|
||||
# [patch."https://github.com/aftershootco/mnn-rs".mnn]
|
||||
# path = "/Users/fs0c131y/Projects/aftershoot/mnn-rs"
|
||||
# [patch."https://github.com/aftershootco/mnn-rs".mnn-sys]
|
||||
# path = "/Users/fs0c131y/Projects/aftershoot/mnn-rs/mnn-sys"
|
||||
# [patch."https://github.com/aftershootco/mnn-rs".mnn-sync]
|
||||
# path = "/Users/fs0c131y/Projects/aftershoot/mnn-rs/mnn-sync"
|
||||
# [patch."https://github.com/aftershootco/mnn-rs".mnn-bridge]
|
||||
# path = "/Users/fs0c131y/Projects/aftershoot/mnn-rs/mnn-bridge"
|
||||
|
||||
[dependencies]
|
||||
clap = { version = "4.5", features = ["derive"] }
|
||||
clap_complete = "4.5"
|
||||
error-stack = "0.5"
|
||||
image = "0.25.6"
|
||||
linfa = "0.7.1"
|
||||
# mnn = { git = "https://github.com/aftershootco/mnn-rs", version = "0.2.0", features = [
|
||||
mnn = { path = "/Users/fs0c131y/Projects/aftershoot/mnn-rs", version = "0.2.0", features = [
|
||||
"metal",
|
||||
"tracing",
|
||||
] }
|
||||
mnn-bridge = { path = "/Users/fs0c131y/Projects/aftershoot/mnn-rs/mnn-bridge", version = "0.1.0", features = [
|
||||
"ndarray",
|
||||
] }
|
||||
mnn-sync = { path = "/Users/fs0c131y/Projects/aftershoot/mnn-rs/mnn-sync", version = "0.1.0", features = [
|
||||
"tracing",
|
||||
] }
|
||||
ndarray = "0.16.1"
|
||||
ndarray-image = { version = "0.1.0", path = "ndarray-image" }
|
||||
rusqlite = { version = "0.36.0", features = ["modern-full"] }
|
||||
thiserror = "2.0"
|
||||
tokio = "1.43.1"
|
||||
tracing = "0.1"
|
||||
tracing-subscriber = { version = "0.3", features = ["env-filter"] }
|
||||
|
||||
[profile.release]
|
||||
debug = true
|
||||
13
bbox/Cargo.toml
Normal file
13
bbox/Cargo.toml
Normal file
@@ -0,0 +1,13 @@
|
||||
[package]
|
||||
name = "bbox"
|
||||
version = "0.1.0"
|
||||
edition = "2024"
|
||||
|
||||
[dependencies]
|
||||
ndarray = "0.16"
|
||||
num = "0.4.3"
|
||||
serde = { version = "1", features = ["derive"], optional = true }
|
||||
|
||||
[features]
|
||||
serde = ["dep:serde"]
|
||||
default = ["serde"]
|
||||
708
bbox/src/lib.rs
Normal file
708
bbox/src/lib.rs
Normal file
@@ -0,0 +1,708 @@
|
||||
pub mod traits;
|
||||
|
||||
/// A bounding box of co-ordinates whose origin is at the top-left corner.
|
||||
#[derive(
|
||||
Debug, Copy, Clone, PartialEq, PartialOrd, Eq, Hash, serde::Serialize, serde::Deserialize,
|
||||
)]
|
||||
#[non_exhaustive]
|
||||
pub struct BBox<T = f32> {
|
||||
pub x: T,
|
||||
pub y: T,
|
||||
pub width: T,
|
||||
pub height: T,
|
||||
}
|
||||
|
||||
impl<T> From<[T; 4]> for BBox<T> {
|
||||
fn from([x, y, width, height]: [T; 4]) -> Self {
|
||||
Self {
|
||||
x,
|
||||
y,
|
||||
width,
|
||||
height,
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
impl<T: Copy> BBox<T> {
|
||||
pub fn new(x: T, y: T, width: T, height: T) -> Self {
|
||||
Self {
|
||||
x,
|
||||
y,
|
||||
width,
|
||||
height,
|
||||
}
|
||||
}
|
||||
|
||||
/// Casts the internal values to another type using [as] keyword
|
||||
pub fn cast<T2>(self) -> BBox<T2>
|
||||
where
|
||||
T: num::cast::AsPrimitive<T2>,
|
||||
T2: Copy + 'static,
|
||||
{
|
||||
BBox {
|
||||
x: self.x.as_(),
|
||||
y: self.y.as_(),
|
||||
width: self.width.as_(),
|
||||
height: self.height.as_(),
|
||||
}
|
||||
}
|
||||
|
||||
/// Clamps all the internal values to the given min and max.
|
||||
pub fn clamp(&self, min: T, max: T) -> Self
|
||||
where
|
||||
T: std::cmp::PartialOrd,
|
||||
{
|
||||
Self {
|
||||
x: num::clamp(self.x, min, max),
|
||||
y: num::clamp(self.y, min, max),
|
||||
width: num::clamp(self.width, min, max),
|
||||
height: num::clamp(self.height, min, max),
|
||||
}
|
||||
}
|
||||
|
||||
pub fn clamp_box(&self, bbox: BBox<T>) -> Self
|
||||
where
|
||||
T: std::cmp::PartialOrd,
|
||||
T: num::Zero,
|
||||
T: core::ops::Add<Output = T>,
|
||||
T: core::ops::Sub<Output = T>,
|
||||
{
|
||||
let x1 = num::clamp(self.x1(), bbox.x1(), bbox.x2());
|
||||
let y1 = num::clamp(self.y1(), bbox.y1(), bbox.y2());
|
||||
let x2 = num::clamp(self.x2(), bbox.x1(), bbox.x2());
|
||||
let y2 = num::clamp(self.y2(), bbox.y1(), bbox.y2());
|
||||
Self::new_xyxy(x1, y1, x2, y2)
|
||||
}
|
||||
|
||||
pub fn normalize(&self, width: T, height: T) -> Self
|
||||
where
|
||||
T: core::ops::Div<Output = T> + Copy,
|
||||
{
|
||||
Self {
|
||||
x: self.x / width,
|
||||
y: self.y / height,
|
||||
width: self.width / width,
|
||||
height: self.height / height,
|
||||
}
|
||||
}
|
||||
|
||||
/// Normalize after casting to float
|
||||
pub fn normalize_f64(&self, width: T, height: T) -> BBox<f64>
|
||||
where
|
||||
T: core::ops::Div<Output = T> + Copy,
|
||||
T: num::cast::AsPrimitive<f64>,
|
||||
{
|
||||
BBox {
|
||||
x: self.x.as_() / width.as_(),
|
||||
y: self.y.as_() / height.as_(),
|
||||
width: self.width.as_() / width.as_(),
|
||||
height: self.height.as_() / height.as_(),
|
||||
}
|
||||
}
|
||||
|
||||
pub fn denormalize(&self, width: T, height: T) -> Self
|
||||
where
|
||||
T: core::ops::Mul<Output = T> + Copy,
|
||||
{
|
||||
Self {
|
||||
x: self.x * width,
|
||||
y: self.y * height,
|
||||
width: self.width * width,
|
||||
height: self.height * height,
|
||||
}
|
||||
}
|
||||
|
||||
pub fn height(&self) -> T {
|
||||
self.height
|
||||
}
|
||||
|
||||
pub fn width(&self) -> T {
|
||||
self.width
|
||||
}
|
||||
|
||||
pub fn padding(&self, padding: T) -> Self
|
||||
where
|
||||
T: core::ops::Add<Output = T> + core::ops::Sub<Output = T> + Copy,
|
||||
{
|
||||
Self {
|
||||
x: self.x - padding,
|
||||
y: self.y - padding,
|
||||
width: self.width + padding + padding,
|
||||
height: self.height + padding + padding,
|
||||
}
|
||||
}
|
||||
|
||||
pub fn padding_height(&self, padding: T) -> Self
|
||||
where
|
||||
T: core::ops::Add<Output = T> + core::ops::Sub<Output = T> + Copy,
|
||||
{
|
||||
Self {
|
||||
x: self.x,
|
||||
y: self.y - padding,
|
||||
width: self.width,
|
||||
height: self.height + padding + padding,
|
||||
}
|
||||
}
|
||||
|
||||
pub fn padding_width(&self, padding: T) -> Self
|
||||
where
|
||||
T: core::ops::Add<Output = T> + core::ops::Sub<Output = T> + Copy,
|
||||
{
|
||||
Self {
|
||||
x: self.x - padding,
|
||||
y: self.y,
|
||||
width: self.width + padding + padding,
|
||||
height: self.height,
|
||||
}
|
||||
}
|
||||
|
||||
// Enlarge / shrink the bounding box by a factor while
|
||||
// keeping the center point and the aspect ratio fixed
|
||||
pub fn scale(&self, factor: T) -> Self
|
||||
where
|
||||
T: core::ops::Mul<Output = T>,
|
||||
T: core::ops::Sub<Output = T>,
|
||||
T: core::ops::Add<Output = T>,
|
||||
T: core::ops::Div<Output = T>,
|
||||
T: num::One + Copy,
|
||||
{
|
||||
let two = num::one::<T>() + num::one::<T>();
|
||||
let width = self.width * factor;
|
||||
let height = self.height * factor;
|
||||
let width_inc = width - self.width;
|
||||
let height_inc = height - self.height;
|
||||
Self {
|
||||
x: self.x - width_inc / two,
|
||||
y: self.y - height_inc / two,
|
||||
width,
|
||||
height,
|
||||
}
|
||||
}
|
||||
|
||||
pub fn scale_x(&self, factor: T) -> Self
|
||||
where
|
||||
T: core::ops::Mul<Output = T>
|
||||
+ core::ops::Sub<Output = T>
|
||||
+ core::ops::Add<Output = T>
|
||||
+ core::ops::Div<Output = T>
|
||||
+ num::One
|
||||
+ Copy,
|
||||
{
|
||||
let two = num::one::<T>() + num::one::<T>();
|
||||
let width = self.width * factor;
|
||||
let width_inc = width - self.width;
|
||||
Self {
|
||||
x: self.x - width_inc / two,
|
||||
y: self.y,
|
||||
width,
|
||||
height: self.height,
|
||||
}
|
||||
}
|
||||
|
||||
pub fn scale_y(&self, factor: T) -> Self
|
||||
where
|
||||
T: core::ops::Mul<Output = T>
|
||||
+ core::ops::Sub<Output = T>
|
||||
+ core::ops::Add<Output = T>
|
||||
+ core::ops::Div<Output = T>
|
||||
+ num::One
|
||||
+ Copy,
|
||||
{
|
||||
let two = num::one::<T>() + num::one::<T>();
|
||||
let height = self.height * factor;
|
||||
let height_inc = height - self.height;
|
||||
Self {
|
||||
x: self.x,
|
||||
y: self.y - height_inc / two,
|
||||
width: self.width,
|
||||
height,
|
||||
}
|
||||
}
|
||||
|
||||
pub fn offset(&self, offset: Point<T>) -> Self
|
||||
where
|
||||
T: core::ops::Add<Output = T> + Copy,
|
||||
{
|
||||
Self {
|
||||
x: self.x + offset.x,
|
||||
y: self.y + offset.y,
|
||||
width: self.width,
|
||||
height: self.height,
|
||||
}
|
||||
}
|
||||
|
||||
/// Translate the bounding box by the given offset
|
||||
/// if they are in the same scale
|
||||
pub fn translate(&self, bbox: Self) -> Self
|
||||
where
|
||||
T: core::ops::Add<Output = T> + Copy,
|
||||
{
|
||||
Self {
|
||||
x: self.x + bbox.x,
|
||||
y: self.y + bbox.y,
|
||||
width: self.width,
|
||||
height: self.height,
|
||||
}
|
||||
}
|
||||
|
||||
pub fn with_top_left(&self, top_left: Point<T>) -> Self {
|
||||
Self {
|
||||
x: top_left.x,
|
||||
y: top_left.y,
|
||||
width: self.width,
|
||||
height: self.height,
|
||||
}
|
||||
}
|
||||
|
||||
pub fn center(&self) -> Point<T>
|
||||
where
|
||||
T: core::ops::Add<Output = T> + core::ops::Div<Output = T> + Copy,
|
||||
T: num::One,
|
||||
{
|
||||
let two = T::one() + T::one();
|
||||
Point::new(self.x + self.width / two, self.y + self.height / two)
|
||||
}
|
||||
|
||||
pub fn area(&self) -> T
|
||||
where
|
||||
T: core::ops::Mul<Output = T> + Copy,
|
||||
{
|
||||
self.width * self.height
|
||||
}
|
||||
|
||||
// Corresponds to self.x1() and self.y1()
|
||||
pub fn top_left(&self) -> Point<T> {
|
||||
Point::new(self.x, self.y)
|
||||
}
|
||||
|
||||
pub fn top_right(&self) -> Point<T>
|
||||
where
|
||||
T: core::ops::Add<Output = T> + Copy,
|
||||
{
|
||||
Point::new(self.x + self.width, self.y)
|
||||
}
|
||||
|
||||
pub fn bottom_left(&self) -> Point<T>
|
||||
where
|
||||
T: core::ops::Add<Output = T> + Copy,
|
||||
{
|
||||
Point::new(self.x, self.y + self.height)
|
||||
}
|
||||
|
||||
// Corresponds to self.x2() and self.y2()
|
||||
pub fn bottom_right(&self) -> Point<T>
|
||||
where
|
||||
T: core::ops::Add<Output = T> + Copy,
|
||||
{
|
||||
Point::new(self.x + self.width, self.y + self.height)
|
||||
}
|
||||
|
||||
pub const fn x1(&self) -> T {
|
||||
self.x
|
||||
}
|
||||
|
||||
pub const fn y1(&self) -> T {
|
||||
self.y
|
||||
}
|
||||
|
||||
pub fn x2(&self) -> T
|
||||
where
|
||||
T: core::ops::Add<Output = T> + Copy,
|
||||
{
|
||||
self.x + self.width
|
||||
}
|
||||
|
||||
pub fn y2(&self) -> T
|
||||
where
|
||||
T: core::ops::Add<Output = T> + Copy,
|
||||
{
|
||||
self.y + self.height
|
||||
}
|
||||
|
||||
pub fn overlap(&self, other: &Self) -> T
|
||||
where
|
||||
T: std::cmp::PartialOrd
|
||||
+ traits::min::Min
|
||||
+ traits::max::Max
|
||||
+ num::Zero
|
||||
+ core::ops::Add<Output = T>
|
||||
+ core::ops::Sub<Output = T>
|
||||
+ core::ops::Mul<Output = T>
|
||||
+ Copy,
|
||||
{
|
||||
let x1 = self.x.max(other.x);
|
||||
let y1 = self.y.max(other.y);
|
||||
let x2 = (self.x + self.width).min(other.x + other.width);
|
||||
let y2 = (self.y + self.height).min(other.y + other.height);
|
||||
let width = (x2 - x1).max(T::zero());
|
||||
let height = (y2 - y1).max(T::zero());
|
||||
width * height
|
||||
}
|
||||
|
||||
pub fn iou(&self, other: &Self) -> T
|
||||
where
|
||||
T: std::cmp::Ord
|
||||
+ num::Zero
|
||||
+ traits::min::Min
|
||||
+ traits::max::Max
|
||||
+ core::ops::Add<Output = T>
|
||||
+ core::ops::Sub<Output = T>
|
||||
+ core::ops::Mul<Output = T>
|
||||
+ core::ops::Div<Output = T>
|
||||
+ Copy,
|
||||
{
|
||||
let overlap = self.overlap(other);
|
||||
let union = self.area() + other.area() - overlap;
|
||||
overlap / union
|
||||
}
|
||||
|
||||
pub fn contains(&self, point: Point<T>) -> bool
|
||||
where
|
||||
T: std::cmp::PartialOrd + core::ops::Add<Output = T> + Copy,
|
||||
{
|
||||
point.x >= self.x
|
||||
&& point.x <= self.x + self.width
|
||||
&& point.y >= self.y
|
||||
&& point.y <= self.y + self.height
|
||||
}
|
||||
|
||||
pub fn contains_bbox(&self, other: Self) -> bool
|
||||
where
|
||||
T: std::cmp::PartialOrd + Copy,
|
||||
T: core::ops::Add<Output = T>,
|
||||
{
|
||||
self.contains(other.top_left())
|
||||
&& self.contains(other.top_right())
|
||||
&& self.contains(other.bottom_left())
|
||||
&& self.contains(other.bottom_right())
|
||||
}
|
||||
|
||||
pub fn new_xywh(x: T, y: T, width: T, height: T) -> Self {
|
||||
Self {
|
||||
x,
|
||||
y,
|
||||
width,
|
||||
height,
|
||||
}
|
||||
}
|
||||
pub fn new_xyxy(x1: T, y1: T, x2: T, y2: T) -> Self
|
||||
where
|
||||
T: core::ops::Sub<Output = T> + Copy,
|
||||
{
|
||||
Self {
|
||||
x: x1,
|
||||
y: y1,
|
||||
width: x2 - x1,
|
||||
height: y2 - y1,
|
||||
}
|
||||
}
|
||||
|
||||
pub fn containing(box1: Self, box2: Self) -> Self
|
||||
where
|
||||
T: traits::min::Min + traits::max::Max + Copy,
|
||||
T: core::ops::Sub<Output = T>,
|
||||
T: core::ops::Add<Output = T>,
|
||||
{
|
||||
let x1 = box1.x.min(box2.x);
|
||||
let y1 = box1.y.min(box2.y);
|
||||
let x2 = box1.x2().max(box2.x2());
|
||||
let y2 = box1.y2().max(box2.y2());
|
||||
Self::new_xyxy(x1, y1, x2, y2)
|
||||
}
|
||||
}
|
||||
|
||||
impl<T: core::ops::Sub<Output = T> + Copy> core::ops::Sub<T> for BBox<T> {
|
||||
type Output = BBox<T>;
|
||||
fn sub(self, rhs: T) -> Self::Output {
|
||||
BBox {
|
||||
x: self.x - rhs,
|
||||
y: self.y - rhs,
|
||||
width: self.width - rhs,
|
||||
height: self.height - rhs,
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
impl<T: core::ops::Add<Output = T> + Copy> core::ops::Add<T> for BBox<T> {
|
||||
type Output = BBox<T>;
|
||||
fn add(self, rhs: T) -> Self::Output {
|
||||
BBox {
|
||||
x: self.x + rhs,
|
||||
y: self.y + rhs,
|
||||
width: self.width + rhs,
|
||||
height: self.height + rhs,
|
||||
}
|
||||
}
|
||||
}
|
||||
impl<T: core::ops::Mul<Output = T> + Copy> core::ops::Mul<T> for BBox<T> {
|
||||
type Output = BBox<T>;
|
||||
fn mul(self, rhs: T) -> Self::Output {
|
||||
BBox {
|
||||
x: self.x * rhs,
|
||||
y: self.y * rhs,
|
||||
width: self.width * rhs,
|
||||
height: self.height * rhs,
|
||||
}
|
||||
}
|
||||
}
|
||||
impl<T: core::ops::Div<Output = T> + Copy> core::ops::Div<T> for BBox<T> {
|
||||
type Output = BBox<T>;
|
||||
fn div(self, rhs: T) -> Self::Output {
|
||||
BBox {
|
||||
x: self.x / rhs,
|
||||
y: self.y / rhs,
|
||||
width: self.width / rhs,
|
||||
height: self.height / rhs,
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
impl<T> core::ops::Add<BBox<T>> for BBox<T>
|
||||
where
|
||||
T: core::ops::Sub<Output = T>
|
||||
+ core::ops::Add<Output = T>
|
||||
+ traits::min::Min
|
||||
+ traits::max::Max
|
||||
+ Copy,
|
||||
{
|
||||
type Output = BBox<T>;
|
||||
fn add(self, rhs: BBox<T>) -> Self::Output {
|
||||
let x1 = self.x1().min(rhs.x1());
|
||||
let y1 = self.y1().min(rhs.y1());
|
||||
let x2 = self.x2().max(rhs.x2());
|
||||
let y2 = self.y2().max(rhs.y2());
|
||||
BBox::new_xyxy(x1, y1, x2, y2)
|
||||
}
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_bbox_add() {
|
||||
let bbox1: BBox<usize> = BBox::new_xyxy(0, 0, 10, 10);
|
||||
let bbox2: BBox<usize> = BBox::new_xyxy(5, 5, 15, 15);
|
||||
let bbox3: BBox<usize> = bbox1 + bbox2;
|
||||
assert_eq!(bbox3, BBox::new_xyxy(0, 0, 15, 15).cast());
|
||||
}
|
||||
|
||||
#[derive(
|
||||
Debug, Copy, Clone, serde::Serialize, serde::Deserialize, PartialEq, PartialOrd, Eq, Ord, Hash,
|
||||
)]
|
||||
pub struct Point<T = f32> {
|
||||
x: T,
|
||||
y: T,
|
||||
}
|
||||
|
||||
impl<T> Point<T> {
|
||||
pub const fn new(x: T, y: T) -> Self {
|
||||
Self { x, y }
|
||||
}
|
||||
|
||||
pub const fn x(&self) -> T
|
||||
where
|
||||
T: Copy,
|
||||
{
|
||||
self.x
|
||||
}
|
||||
|
||||
pub const fn y(&self) -> T
|
||||
where
|
||||
T: Copy,
|
||||
{
|
||||
self.y
|
||||
}
|
||||
|
||||
pub fn cast<T2>(&self) -> Point<T2>
|
||||
where
|
||||
T: num::cast::AsPrimitive<T2>,
|
||||
T2: Copy + 'static,
|
||||
{
|
||||
Point {
|
||||
x: self.x.as_(),
|
||||
y: self.y.as_(),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
impl<T: core::ops::Sub<T, Output = T> + Copy> core::ops::Sub<Point<T>> for Point<T> {
|
||||
type Output = Point<T>;
|
||||
fn sub(self, rhs: Point<T>) -> Self::Output {
|
||||
Point {
|
||||
x: self.x - rhs.x,
|
||||
y: self.y - rhs.y,
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
impl<T: core::ops::Add<T, Output = T> + Copy> core::ops::Add<Point<T>> for Point<T> {
|
||||
type Output = Point<T>;
|
||||
fn add(self, rhs: Point<T>) -> Self::Output {
|
||||
Point {
|
||||
x: self.x + rhs.x,
|
||||
y: self.y + rhs.y,
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
impl<T: core::ops::Sub<Output = T> + Copy> Point<T> {
|
||||
/// If both the boxes are in the same scale then make the translation of the origin to the
|
||||
/// other box
|
||||
pub fn with_origin(&self, origin: Self) -> Self {
|
||||
*self - origin
|
||||
}
|
||||
}
|
||||
|
||||
impl<T: core::ops::Add<Output = T> + Copy> Point<T> {
|
||||
pub fn translate(&self, point: Point<T>) -> Self {
|
||||
*self + point
|
||||
}
|
||||
}
|
||||
|
||||
impl<I: num::Zero> BBox<I>
|
||||
where
|
||||
I: num::cast::AsPrimitive<usize>,
|
||||
{
|
||||
pub fn zeros_ndarray_2d<T: num::Zero + Copy>(&self) -> ndarray::Array2<T> {
|
||||
ndarray::Array2::<T>::zeros((self.height.as_(), self.width.as_()))
|
||||
}
|
||||
pub fn zeros_ndarray_3d<T: num::Zero + Copy>(&self, channels: usize) -> ndarray::Array3<T> {
|
||||
ndarray::Array3::<T>::zeros((self.height.as_(), self.width.as_(), channels))
|
||||
}
|
||||
pub fn ones_ndarray_2d<T: num::One + Copy>(&self) -> ndarray::Array2<T> {
|
||||
ndarray::Array2::<T>::ones((self.height.as_(), self.width.as_()))
|
||||
}
|
||||
}
|
||||
|
||||
impl<T: num::Float> BBox<T> {
|
||||
pub fn round(&self) -> Self {
|
||||
Self {
|
||||
x: self.x.round(),
|
||||
y: self.y.round(),
|
||||
width: self.width.round(),
|
||||
height: self.height.round(),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
#[cfg(test)]
|
||||
mod bbox_clamp_tests {
|
||||
use super::*;
|
||||
#[test]
|
||||
pub fn bbox_test_clamp_box() {
|
||||
let large_box = BBox::new(0, 0, 100, 100);
|
||||
let small_box = BBox::new(10, 10, 20, 20);
|
||||
let clamped = large_box.clamp_box(small_box);
|
||||
assert_eq!(clamped, small_box);
|
||||
}
|
||||
|
||||
#[test]
|
||||
pub fn bbox_test_clamp_box_offset() {
|
||||
let box_a = BBox::new(0, 0, 100, 100);
|
||||
let box_b = BBox::new(-10, -10, 20, 20);
|
||||
let clamped = box_b.clamp_box(box_a);
|
||||
let expected = BBox::new(0, 0, 10, 10);
|
||||
assert_eq!(expected, clamped);
|
||||
}
|
||||
}
|
||||
|
||||
#[cfg(test)]
|
||||
mod bbox_padding_tests {
|
||||
use super::*;
|
||||
#[test]
|
||||
pub fn bbox_test_padding() {
|
||||
let bbox = BBox::new(0, 0, 10, 10);
|
||||
let padded = bbox.padding(2);
|
||||
assert_eq!(padded, BBox::new(-2, -2, 14, 14));
|
||||
}
|
||||
|
||||
#[test]
|
||||
pub fn bbox_test_padding_height() {
|
||||
let bbox = BBox::new(0, 0, 10, 10);
|
||||
let padded = bbox.padding_height(2);
|
||||
assert_eq!(padded, BBox::new(0, -2, 10, 14));
|
||||
}
|
||||
|
||||
#[test]
|
||||
pub fn bbox_test_padding_width() {
|
||||
let bbox = BBox::new(0, 0, 10, 10);
|
||||
let padded = bbox.padding_width(2);
|
||||
assert_eq!(padded, BBox::new(-2, 0, 14, 10));
|
||||
}
|
||||
|
||||
#[test]
|
||||
pub fn bbox_test_clamped_padding() {
|
||||
let bbox = BBox::new(0, 0, 10, 10);
|
||||
let padded = bbox.padding(2);
|
||||
let clamp = BBox::new(0, 0, 12, 12);
|
||||
let clamped = padded.clamp_box(clamp);
|
||||
assert_eq!(clamped, clamp);
|
||||
}
|
||||
|
||||
#[test]
|
||||
pub fn bbox_clamp_failure() {
|
||||
let og = BBox::new(475.0, 79.625, 37.0, 282.15);
|
||||
let padded = BBox {
|
||||
x: 471.3,
|
||||
y: 51.412499999999994,
|
||||
width: 40.69999999999999,
|
||||
height: 338.54999999999995,
|
||||
};
|
||||
let clamp = BBox::new(0.0, 0.0, 512.0, 512.0);
|
||||
let sus = padded.clamp_box(clamp);
|
||||
assert!(clamp.contains_bbox(sus));
|
||||
}
|
||||
}
|
||||
|
||||
#[cfg(test)]
|
||||
mod bbox_scale_tests {
|
||||
use super::*;
|
||||
#[test]
|
||||
pub fn bbox_test_scale_int() {
|
||||
let bbox = BBox::new(0, 0, 10, 10);
|
||||
let scaled = bbox.scale(2);
|
||||
assert_eq!(scaled, BBox::new(-5, -5, 20, 20));
|
||||
}
|
||||
|
||||
#[test]
|
||||
pub fn bbox_test_scale_float() {
|
||||
let bbox = BBox::new(0, 0, 10, 10).cast();
|
||||
let scaled = bbox.scale(1.05); // 5% increase
|
||||
let l = 10.0 * 0.05;
|
||||
assert_eq!(scaled, BBox::new(-l / 2.0, -l / 2.0, 10.0 + l, 10.0 + l));
|
||||
}
|
||||
|
||||
#[test]
|
||||
pub fn bbox_test_scale_float_negative() {
|
||||
let bbox = BBox::new(0, 0, 10, 10).cast();
|
||||
let scaled = bbox.scale(0.95); // 5% decrease
|
||||
let l = -10.0 * 0.05;
|
||||
assert_eq!(scaled, BBox::new(-l / 2.0, -l / 2.0, 10.0 + l, 10.0 + l));
|
||||
}
|
||||
|
||||
#[test]
|
||||
pub fn bbox_scale_float() {
|
||||
let bbox = BBox::new_xywh(0, 0, 200, 200);
|
||||
let scaled = bbox.cast::<f64>().scale(1.1).cast::<i32>().clamp(0, 1000);
|
||||
let expected = BBox::new(0, 0, 220, 220);
|
||||
assert_eq!(scaled, expected);
|
||||
}
|
||||
#[test]
|
||||
pub fn add_padding_bbox_example() {
|
||||
// let result = add_padding_bbox(
|
||||
// vec![Rect::new(100, 200, 300, 400)],
|
||||
// (0.1, 0.1),
|
||||
// (1000, 1000),
|
||||
// );
|
||||
// assert_eq!(result[0], Rect::new(70, 160, 360, 480));
|
||||
let bbox = BBox::new(100, 200, 300, 400);
|
||||
let scaled = bbox.cast::<f64>().scale(1.2).cast::<i32>().clamp(0, 1000);
|
||||
assert_eq!(bbox, BBox::new(100, 200, 300, 400));
|
||||
assert_eq!(scaled, BBox::new(70, 160, 360, 480));
|
||||
}
|
||||
#[test]
|
||||
pub fn scale_bboxes() {
|
||||
// let result = scale_bboxes(Rect::new(100, 200, 300, 400), (1000, 1000), (500, 500));
|
||||
// assert_eq!(result[0], Rect::new(200, 400, 600, 800));
|
||||
let bbox = BBox::new(100, 200, 300, 400);
|
||||
let scaled = bbox.scale(2);
|
||||
assert_eq!(scaled, BBox::new(200, 400, 600, 800));
|
||||
}
|
||||
}
|
||||
2
bbox/src/traits.rs
Normal file
2
bbox/src/traits.rs
Normal file
@@ -0,0 +1,2 @@
|
||||
pub mod max;
|
||||
pub mod min;
|
||||
27
bbox/src/traits/max.rs
Normal file
27
bbox/src/traits/max.rs
Normal file
@@ -0,0 +1,27 @@
|
||||
pub trait Max: Sized + Copy {
|
||||
fn max(self, other: Self) -> Self;
|
||||
}
|
||||
|
||||
macro_rules! impl_max {
|
||||
($($t:ty),*) => {
|
||||
$(
|
||||
impl Max for $t {
|
||||
fn max(self, other: Self) -> Self {
|
||||
Ord::max(self, other)
|
||||
}
|
||||
}
|
||||
)*
|
||||
};
|
||||
(float $($t:ty),*) => {
|
||||
$(
|
||||
impl Max for $t {
|
||||
fn max(self, other: Self) -> Self {
|
||||
Self::max(self, other)
|
||||
}
|
||||
}
|
||||
)*
|
||||
};
|
||||
}
|
||||
|
||||
impl_max!(usize, u8, u16, u32, u64, u128, isize, i8, i16, i32, i64, i128);
|
||||
impl_max!(float f32, f64);
|
||||
27
bbox/src/traits/min.rs
Normal file
27
bbox/src/traits/min.rs
Normal file
@@ -0,0 +1,27 @@
|
||||
pub trait Min: Sized + Copy {
|
||||
fn min(self, other: Self) -> Self;
|
||||
}
|
||||
|
||||
macro_rules! impl_min {
|
||||
($($t:ty),*) => {
|
||||
$(
|
||||
impl Min for $t {
|
||||
fn min(self, other: Self) -> Self {
|
||||
Ord::min(self, other)
|
||||
}
|
||||
}
|
||||
)*
|
||||
};
|
||||
(float $($t:ty),*) => {
|
||||
$(
|
||||
impl Min for $t {
|
||||
fn min(self, other: Self) -> Self {
|
||||
Self::min(self, other)
|
||||
}
|
||||
}
|
||||
)*
|
||||
};
|
||||
}
|
||||
|
||||
impl_min!(usize, u8, u16, u32, u64, u128, isize, i8, i16, i32, i64, i128);
|
||||
impl_min!(float f32, f64);
|
||||
236
deny.toml
Normal file
236
deny.toml
Normal file
@@ -0,0 +1,236 @@
|
||||
# This template contains all of the possible sections and their default values
|
||||
|
||||
# Note that all fields that take a lint level have these possible values:
|
||||
# * deny - An error will be produced and the check will fail
|
||||
# * warn - A warning will be produced, but the check will not fail
|
||||
# * allow - No warning or error will be produced, though in some cases a note
|
||||
# will be
|
||||
|
||||
# The values provided in this template are the default values that will be used
|
||||
# when any section or field is not specified in your own configuration
|
||||
|
||||
# Root options
|
||||
|
||||
# The graph table configures how the dependency graph is constructed and thus
|
||||
# which crates the checks are performed against
|
||||
[graph]
|
||||
# If 1 or more target triples (and optionally, target_features) are specified,
|
||||
# only the specified targets will be checked when running `cargo deny check`.
|
||||
# This means, if a particular package is only ever used as a target specific
|
||||
# dependency, such as, for example, the `nix` crate only being used via the
|
||||
# `target_family = "unix"` configuration, that only having windows targets in
|
||||
# this list would mean the nix crate, as well as any of its exclusive
|
||||
# dependencies not shared by any other crates, would be ignored, as the target
|
||||
# list here is effectively saying which targets you are building for.
|
||||
targets = [
|
||||
# The triple can be any string, but only the target triples built in to
|
||||
# rustc (as of 1.40) can be checked against actual config expressions
|
||||
#"x86_64-unknown-linux-musl",
|
||||
# You can also specify which target_features you promise are enabled for a
|
||||
# particular target. target_features are currently not validated against
|
||||
# the actual valid features supported by the target architecture.
|
||||
#{ triple = "wasm32-unknown-unknown", features = ["atomics"] },
|
||||
]
|
||||
# When creating the dependency graph used as the source of truth when checks are
|
||||
# executed, this field can be used to prune crates from the graph, removing them
|
||||
# from the view of cargo-deny. This is an extremely heavy hammer, as if a crate
|
||||
# is pruned from the graph, all of its dependencies will also be pruned unless
|
||||
# they are connected to another crate in the graph that hasn't been pruned,
|
||||
# so it should be used with care. The identifiers are [Package ID Specifications]
|
||||
# (https://doc.rust-lang.org/cargo/reference/pkgid-spec.html)
|
||||
#exclude = []
|
||||
# If true, metadata will be collected with `--all-features`. Note that this can't
|
||||
# be toggled off if true, if you want to conditionally enable `--all-features` it
|
||||
# is recommended to pass `--all-features` on the cmd line instead
|
||||
all-features = false
|
||||
# If true, metadata will be collected with `--no-default-features`. The same
|
||||
# caveat with `all-features` applies
|
||||
no-default-features = false
|
||||
# If set, these feature will be enabled when collecting metadata. If `--features`
|
||||
# is specified on the cmd line they will take precedence over this option.
|
||||
#features = []
|
||||
|
||||
# The output table provides options for how/if diagnostics are outputted
|
||||
[output]
|
||||
# When outputting inclusion graphs in diagnostics that include features, this
|
||||
# option can be used to specify the depth at which feature edges will be added.
|
||||
# This option is included since the graphs can be quite large and the addition
|
||||
# of features from the crate(s) to all of the graph roots can be far too verbose.
|
||||
# This option can be overridden via `--feature-depth` on the cmd line
|
||||
feature-depth = 1
|
||||
|
||||
# This section is considered when running `cargo deny check advisories`
|
||||
# More documentation for the advisories section can be found here:
|
||||
# https://embarkstudios.github.io/cargo-deny/checks/advisories/cfg.html
|
||||
[advisories]
|
||||
# The path where the advisory databases are cloned/fetched into
|
||||
#db-path = "$CARGO_HOME/advisory-dbs"
|
||||
# The url(s) of the advisory databases to use
|
||||
#db-urls = ["https://github.com/rustsec/advisory-db"]
|
||||
# A list of advisory IDs to ignore. Note that ignored advisories will still
|
||||
# output a note when they are encountered.
|
||||
ignore = [
|
||||
#"RUSTSEC-0000-0000",
|
||||
#{ id = "RUSTSEC-0000-0000", reason = "you can specify a reason the advisory is ignored" },
|
||||
#"a-crate-that-is-yanked@0.1.1", # you can also ignore yanked crate versions if you wish
|
||||
#{ crate = "a-crate-that-is-yanked@0.1.1", reason = "you can specify why you are ignoring the yanked crate" },
|
||||
]
|
||||
# If this is true, then cargo deny will use the git executable to fetch advisory database.
|
||||
# If this is false, then it uses a built-in git library.
|
||||
# Setting this to true can be helpful if you have special authentication requirements that cargo-deny does not support.
|
||||
# See Git Authentication for more information about setting up git authentication.
|
||||
#git-fetch-with-cli = true
|
||||
|
||||
# This section is considered when running `cargo deny check licenses`
|
||||
# More documentation for the licenses section can be found here:
|
||||
# https://embarkstudios.github.io/cargo-deny/checks/licenses/cfg.html
|
||||
[licenses]
|
||||
# List of explicitly allowed licenses
|
||||
# See https://spdx.org/licenses/ for list of possible licenses
|
||||
# [possible values: any SPDX 3.11 short identifier (+ optional exception)].
|
||||
allow = [
|
||||
"MIT",
|
||||
"Apache-2.0",
|
||||
"Unicode-3.0",
|
||||
#"Apache-2.0 WITH LLVM-exception",
|
||||
]
|
||||
# The confidence threshold for detecting a license from license text.
|
||||
# The higher the value, the more closely the license text must be to the
|
||||
# canonical license text of a valid SPDX license file.
|
||||
# [possible values: any between 0.0 and 1.0].
|
||||
confidence-threshold = 0.8
|
||||
# Allow 1 or more licenses on a per-crate basis, so that particular licenses
|
||||
# aren't accepted for every possible crate as with the normal allow list
|
||||
exceptions = [
|
||||
# Each entry is the crate and version constraint, and its specific allow
|
||||
# list
|
||||
#{ allow = ["Zlib"], crate = "adler32" },
|
||||
]
|
||||
|
||||
# Some crates don't have (easily) machine readable licensing information,
|
||||
# adding a clarification entry for it allows you to manually specify the
|
||||
# licensing information
|
||||
#[[licenses.clarify]]
|
||||
# The package spec the clarification applies to
|
||||
#crate = "ring"
|
||||
# The SPDX expression for the license requirements of the crate
|
||||
#expression = "MIT AND ISC AND OpenSSL"
|
||||
# One or more files in the crate's source used as the "source of truth" for
|
||||
# the license expression. If the contents match, the clarification will be used
|
||||
# when running the license check, otherwise the clarification will be ignored
|
||||
# and the crate will be checked normally, which may produce warnings or errors
|
||||
# depending on the rest of your configuration
|
||||
#license-files = [
|
||||
# Each entry is a crate relative path, and the (opaque) hash of its contents
|
||||
#{ path = "LICENSE", hash = 0xbd0eed23 }
|
||||
#]
|
||||
|
||||
[licenses.private]
|
||||
# If true, ignores workspace crates that aren't published, or are only
|
||||
# published to private registries.
|
||||
# To see how to mark a crate as unpublished (to the official registry),
|
||||
# visit https://doc.rust-lang.org/cargo/reference/manifest.html#the-publish-field.
|
||||
ignore = false
|
||||
# One or more private registries that you might publish crates to, if a crate
|
||||
# is only published to private registries, and ignore is true, the crate will
|
||||
# not have its license(s) checked
|
||||
registries = [
|
||||
#"https://sekretz.com/registry
|
||||
]
|
||||
|
||||
# This section is considered when running `cargo deny check bans`.
|
||||
# More documentation about the 'bans' section can be found here:
|
||||
# https://embarkstudios.github.io/cargo-deny/checks/bans/cfg.html
|
||||
[bans]
|
||||
# Lint level for when multiple versions of the same crate are detected
|
||||
multiple-versions = "warn"
|
||||
# Lint level for when a crate version requirement is `*`
|
||||
wildcards = "allow"
|
||||
# The graph highlighting used when creating dotgraphs for crates
|
||||
# with multiple versions
|
||||
# * lowest-version - The path to the lowest versioned duplicate is highlighted
|
||||
# * simplest-path - The path to the version with the fewest edges is highlighted
|
||||
# * all - Both lowest-version and simplest-path are used
|
||||
highlight = "all"
|
||||
# The default lint level for `default` features for crates that are members of
|
||||
# the workspace that is being checked. This can be overridden by allowing/denying
|
||||
# `default` on a crate-by-crate basis if desired.
|
||||
workspace-default-features = "allow"
|
||||
# The default lint level for `default` features for external crates that are not
|
||||
# members of the workspace. This can be overridden by allowing/denying `default`
|
||||
# on a crate-by-crate basis if desired.
|
||||
external-default-features = "allow"
|
||||
# List of crates that are allowed. Use with care!
|
||||
allow = [
|
||||
#"ansi_term@0.11.0",
|
||||
#{ crate = "ansi_term@0.11.0", reason = "you can specify a reason it is allowed" },
|
||||
]
|
||||
# List of crates to deny
|
||||
deny = [
|
||||
#"ansi_term@0.11.0",
|
||||
#{ crate = "ansi_term@0.11.0", reason = "you can specify a reason it is banned" },
|
||||
# Wrapper crates can optionally be specified to allow the crate when it
|
||||
# is a direct dependency of the otherwise banned crate
|
||||
#{ crate = "ansi_term@0.11.0", wrappers = ["this-crate-directly-depends-on-ansi_term"] },
|
||||
]
|
||||
|
||||
# List of features to allow/deny
|
||||
# Each entry the name of a crate and a version range. If version is
|
||||
# not specified, all versions will be matched.
|
||||
#[[bans.features]]
|
||||
#crate = "reqwest"
|
||||
# Features to not allow
|
||||
#deny = ["json"]
|
||||
# Features to allow
|
||||
#allow = [
|
||||
# "rustls",
|
||||
# "__rustls",
|
||||
# "__tls",
|
||||
# "hyper-rustls",
|
||||
# "rustls",
|
||||
# "rustls-pemfile",
|
||||
# "rustls-tls-webpki-roots",
|
||||
# "tokio-rustls",
|
||||
# "webpki-roots",
|
||||
#]
|
||||
# If true, the allowed features must exactly match the enabled feature set. If
|
||||
# this is set there is no point setting `deny`
|
||||
#exact = true
|
||||
|
||||
# Certain crates/versions that will be skipped when doing duplicate detection.
|
||||
skip = [
|
||||
#"ansi_term@0.11.0",
|
||||
#{ crate = "ansi_term@0.11.0", reason = "you can specify a reason why it can't be updated/removed" },
|
||||
]
|
||||
# Similarly to `skip` allows you to skip certain crates during duplicate
|
||||
# detection. Unlike skip, it also includes the entire tree of transitive
|
||||
# dependencies starting at the specified crate, up to a certain depth, which is
|
||||
# by default infinite.
|
||||
skip-tree = [
|
||||
#"ansi_term@0.11.0", # will be skipped along with _all_ of its direct and transitive dependencies
|
||||
#{ crate = "ansi_term@0.11.0", depth = 20 },
|
||||
]
|
||||
|
||||
# This section is considered when running `cargo deny check sources`.
|
||||
# More documentation about the 'sources' section can be found here:
|
||||
# https://embarkstudios.github.io/cargo-deny/checks/sources/cfg.html
|
||||
[sources]
|
||||
# Lint level for what to happen when a crate from a crate registry that is not
|
||||
# in the allow list is encountered
|
||||
unknown-registry = "warn"
|
||||
# Lint level for what to happen when a crate from a git repository that is not
|
||||
# in the allow list is encountered
|
||||
unknown-git = "warn"
|
||||
# List of URLs for allowed crate registries. Defaults to the crates.io index
|
||||
# if not specified. If it is specified but empty, no registries are allowed.
|
||||
allow-registry = ["https://github.com/rust-lang/crates.io-index"]
|
||||
# List of URLs for allowed Git repositories
|
||||
allow-git = []
|
||||
|
||||
[sources.allow-org]
|
||||
# github.com organizations to allow git sources for
|
||||
github = []
|
||||
# gitlab.com organizations to allow git sources for
|
||||
gitlab = []
|
||||
# bitbucket.org organizations to allow git sources for
|
||||
bitbucket = []
|
||||
227
flake.lock
generated
Normal file
227
flake.lock
generated
Normal file
@@ -0,0 +1,227 @@
|
||||
{
|
||||
"nodes": {
|
||||
"advisory-db": {
|
||||
"flake": false,
|
||||
"locked": {
|
||||
"lastModified": 1750151065,
|
||||
"narHash": "sha256-il+CAqChFIB82xP6bO43dWlUVs+NlG7a4g8liIP5HcI=",
|
||||
"owner": "rustsec",
|
||||
"repo": "advisory-db",
|
||||
"rev": "7573f55ba337263f61167dbb0ea926cdc7c8eb5d",
|
||||
"type": "github"
|
||||
},
|
||||
"original": {
|
||||
"owner": "rustsec",
|
||||
"repo": "advisory-db",
|
||||
"type": "github"
|
||||
}
|
||||
},
|
||||
"crane": {
|
||||
"locked": {
|
||||
"lastModified": 1750266157,
|
||||
"narHash": "sha256-tL42YoNg9y30u7zAqtoGDNdTyXTi8EALDeCB13FtbQA=",
|
||||
"owner": "ipetkov",
|
||||
"repo": "crane",
|
||||
"rev": "e37c943371b73ed87faf33f7583860f81f1d5a48",
|
||||
"type": "github"
|
||||
},
|
||||
"original": {
|
||||
"owner": "ipetkov",
|
||||
"repo": "crane",
|
||||
"type": "github"
|
||||
}
|
||||
},
|
||||
"flake-utils": {
|
||||
"inputs": {
|
||||
"systems": "systems"
|
||||
},
|
||||
"locked": {
|
||||
"lastModified": 1731533236,
|
||||
"narHash": "sha256-l0KFg5HjrsfsO/JpG+r7fRrqm12kzFHyUHqHCVpMMbI=",
|
||||
"owner": "numtide",
|
||||
"repo": "flake-utils",
|
||||
"rev": "11707dc2f618dd54ca8739b309ec4fc024de578b",
|
||||
"type": "github"
|
||||
},
|
||||
"original": {
|
||||
"owner": "numtide",
|
||||
"repo": "flake-utils",
|
||||
"type": "github"
|
||||
}
|
||||
},
|
||||
"flake-utils_2": {
|
||||
"inputs": {
|
||||
"systems": "systems_2"
|
||||
},
|
||||
"locked": {
|
||||
"lastModified": 1731533236,
|
||||
"narHash": "sha256-l0KFg5HjrsfsO/JpG+r7fRrqm12kzFHyUHqHCVpMMbI=",
|
||||
"owner": "numtide",
|
||||
"repo": "flake-utils",
|
||||
"rev": "11707dc2f618dd54ca8739b309ec4fc024de578b",
|
||||
"type": "github"
|
||||
},
|
||||
"original": {
|
||||
"owner": "numtide",
|
||||
"repo": "flake-utils",
|
||||
"type": "github"
|
||||
}
|
||||
},
|
||||
"mnn": {
|
||||
"flake": false,
|
||||
"locked": {
|
||||
"lastModified": 1749173738,
|
||||
"narHash": "sha256-pNljvQ4xMZ4VmuxQyXt+boNBZD0+UZNpNLrWrj8Rtfw=",
|
||||
"owner": "alibaba",
|
||||
"repo": "MNN",
|
||||
"rev": "ebdada82634300956e08bd4056ecfeb1e4f23b32",
|
||||
"type": "github"
|
||||
},
|
||||
"original": {
|
||||
"owner": "alibaba",
|
||||
"ref": "3.2.0",
|
||||
"repo": "MNN",
|
||||
"type": "github"
|
||||
}
|
||||
},
|
||||
"mnn-overlay": {
|
||||
"inputs": {
|
||||
"flake-utils": "flake-utils_2",
|
||||
"mnn": "mnn",
|
||||
"nixpkgs": [
|
||||
"nixpkgs"
|
||||
]
|
||||
},
|
||||
"locked": {
|
||||
"lastModified": 1749204972,
|
||||
"narHash": "sha256-ICLU408iwxZA7uETBmEBuuForBIPLvonuy1hW/fuiME=",
|
||||
"owner": "uttarayan21",
|
||||
"repo": "mnn-nix-overlay",
|
||||
"rev": "7b97393977689e851a6840a8e1cbea058e67363a",
|
||||
"type": "github"
|
||||
},
|
||||
"original": {
|
||||
"owner": "uttarayan21",
|
||||
"repo": "mnn-nix-overlay",
|
||||
"type": "github"
|
||||
}
|
||||
},
|
||||
"mnn-src": {
|
||||
"flake": false,
|
||||
"locked": {
|
||||
"lastModified": 1749173738,
|
||||
"narHash": "sha256-pNljvQ4xMZ4VmuxQyXt+boNBZD0+UZNpNLrWrj8Rtfw=",
|
||||
"owner": "alibaba",
|
||||
"repo": "MNN",
|
||||
"rev": "ebdada82634300956e08bd4056ecfeb1e4f23b32",
|
||||
"type": "github"
|
||||
},
|
||||
"original": {
|
||||
"owner": "alibaba",
|
||||
"ref": "3.2.0",
|
||||
"repo": "MNN",
|
||||
"type": "github"
|
||||
}
|
||||
},
|
||||
"nix-github-actions": {
|
||||
"inputs": {
|
||||
"nixpkgs": [
|
||||
"nixpkgs"
|
||||
]
|
||||
},
|
||||
"locked": {
|
||||
"lastModified": 1737420293,
|
||||
"narHash": "sha256-F1G5ifvqTpJq7fdkT34e/Jy9VCyzd5XfJ9TO8fHhJWE=",
|
||||
"owner": "nix-community",
|
||||
"repo": "nix-github-actions",
|
||||
"rev": "f4158fa080ef4503c8f4c820967d946c2af31ec9",
|
||||
"type": "github"
|
||||
},
|
||||
"original": {
|
||||
"owner": "nix-community",
|
||||
"repo": "nix-github-actions",
|
||||
"type": "github"
|
||||
}
|
||||
},
|
||||
"nixpkgs": {
|
||||
"locked": {
|
||||
"lastModified": 1750506804,
|
||||
"narHash": "sha256-VLFNc4egNjovYVxDGyBYTrvVCgDYgENp5bVi9fPTDYc=",
|
||||
"owner": "nixos",
|
||||
"repo": "nixpkgs",
|
||||
"rev": "4206c4cb56751df534751b058295ea61357bbbaa",
|
||||
"type": "github"
|
||||
},
|
||||
"original": {
|
||||
"owner": "nixos",
|
||||
"ref": "nixos-unstable",
|
||||
"repo": "nixpkgs",
|
||||
"type": "github"
|
||||
}
|
||||
},
|
||||
"root": {
|
||||
"inputs": {
|
||||
"advisory-db": "advisory-db",
|
||||
"crane": "crane",
|
||||
"flake-utils": "flake-utils",
|
||||
"mnn-overlay": "mnn-overlay",
|
||||
"mnn-src": "mnn-src",
|
||||
"nix-github-actions": "nix-github-actions",
|
||||
"nixpkgs": "nixpkgs",
|
||||
"rust-overlay": "rust-overlay"
|
||||
}
|
||||
},
|
||||
"rust-overlay": {
|
||||
"inputs": {
|
||||
"nixpkgs": [
|
||||
"nixpkgs"
|
||||
]
|
||||
},
|
||||
"locked": {
|
||||
"lastModified": 1750732748,
|
||||
"narHash": "sha256-HR2b3RHsPeJm+Fb+1ui8nXibgniVj7hBNvUbXEyz0DU=",
|
||||
"owner": "oxalica",
|
||||
"repo": "rust-overlay",
|
||||
"rev": "4b4494b2ba7e8a8041b2e28320b2ee02c115c75f",
|
||||
"type": "github"
|
||||
},
|
||||
"original": {
|
||||
"owner": "oxalica",
|
||||
"repo": "rust-overlay",
|
||||
"type": "github"
|
||||
}
|
||||
},
|
||||
"systems": {
|
||||
"locked": {
|
||||
"lastModified": 1681028828,
|
||||
"narHash": "sha256-Vy1rq5AaRuLzOxct8nz4T6wlgyUR7zLU309k9mBC768=",
|
||||
"owner": "nix-systems",
|
||||
"repo": "default",
|
||||
"rev": "da67096a3b9bf56a91d16901293e51ba5b49a27e",
|
||||
"type": "github"
|
||||
},
|
||||
"original": {
|
||||
"owner": "nix-systems",
|
||||
"repo": "default",
|
||||
"type": "github"
|
||||
}
|
||||
},
|
||||
"systems_2": {
|
||||
"locked": {
|
||||
"lastModified": 1681028828,
|
||||
"narHash": "sha256-Vy1rq5AaRuLzOxct8nz4T6wlgyUR7zLU309k9mBC768=",
|
||||
"owner": "nix-systems",
|
||||
"repo": "default",
|
||||
"rev": "da67096a3b9bf56a91d16901293e51ba5b49a27e",
|
||||
"type": "github"
|
||||
},
|
||||
"original": {
|
||||
"owner": "nix-systems",
|
||||
"repo": "default",
|
||||
"type": "github"
|
||||
}
|
||||
}
|
||||
},
|
||||
"root": "root",
|
||||
"version": 7
|
||||
}
|
||||
176
flake.nix
Normal file
176
flake.nix
Normal file
@@ -0,0 +1,176 @@
|
||||
{
|
||||
description = "A simple rust flake using rust-overlay and craneLib";
|
||||
|
||||
inputs = {
|
||||
nixpkgs.url = "github:nixos/nixpkgs/nixos-unstable";
|
||||
flake-utils.url = "github:numtide/flake-utils";
|
||||
crane.url = "github:ipetkov/crane";
|
||||
nix-github-actions = {
|
||||
url = "github:nix-community/nix-github-actions";
|
||||
inputs.nixpkgs.follows = "nixpkgs";
|
||||
};
|
||||
rust-overlay = {
|
||||
url = "github:oxalica/rust-overlay";
|
||||
inputs.nixpkgs.follows = "nixpkgs";
|
||||
};
|
||||
advisory-db = {
|
||||
url = "github:rustsec/advisory-db";
|
||||
flake = false;
|
||||
};
|
||||
mnn-overlay = {
|
||||
url = "github:uttarayan21/mnn-nix-overlay";
|
||||
inputs.nixpkgs.follows = "nixpkgs";
|
||||
};
|
||||
mnn-src = {
|
||||
url = "github:alibaba/MNN/3.2.0";
|
||||
flake = false;
|
||||
};
|
||||
};
|
||||
|
||||
outputs = {
|
||||
self,
|
||||
crane,
|
||||
flake-utils,
|
||||
nixpkgs,
|
||||
rust-overlay,
|
||||
advisory-db,
|
||||
nix-github-actions,
|
||||
mnn-overlay,
|
||||
mnn-src,
|
||||
...
|
||||
}:
|
||||
flake-utils.lib.eachDefaultSystem (
|
||||
system: let
|
||||
pkgs = import nixpkgs {
|
||||
inherit system;
|
||||
overlays = [
|
||||
rust-overlay.overlays.default
|
||||
(final: prev: {
|
||||
mnn = mnn-overlay.packages.${system}.mnn.override {
|
||||
src = mnn-src;
|
||||
buildConverter = true;
|
||||
enableMetal = true;
|
||||
enableOpencl = true;
|
||||
};
|
||||
})
|
||||
];
|
||||
};
|
||||
inherit (pkgs) lib;
|
||||
cargoToml = builtins.fromTOML (builtins.readFile ./Cargo.toml);
|
||||
name = cargoToml.package.name;
|
||||
|
||||
stableToolchain = pkgs.rust-bin.stable.latest.default;
|
||||
stableToolchainWithLLvmTools = stableToolchain.override {
|
||||
extensions = ["rust-src" "llvm-tools"];
|
||||
};
|
||||
stableToolchainWithRustAnalyzer = stableToolchain.override {
|
||||
extensions = ["rust-src" "rust-analyzer"];
|
||||
};
|
||||
craneLib = (crane.mkLib pkgs).overrideToolchain stableToolchain;
|
||||
craneLibLLvmTools = (crane.mkLib pkgs).overrideToolchain stableToolchainWithLLvmTools;
|
||||
|
||||
src = let
|
||||
filterBySuffix = path: exts: lib.any (ext: lib.hasSuffix ext path) exts;
|
||||
sourceFilters = path: type: (craneLib.filterCargoSources path type) || filterBySuffix path [".c" ".h" ".hpp" ".cpp" ".cc"];
|
||||
in
|
||||
lib.cleanSourceWith {
|
||||
filter = sourceFilters;
|
||||
src = ./.;
|
||||
};
|
||||
commonArgs =
|
||||
{
|
||||
inherit src;
|
||||
pname = name;
|
||||
stdenv = pkgs.clangStdenv;
|
||||
doCheck = false;
|
||||
# LIBCLANG_PATH = "${pkgs.llvmPackages.libclang.lib}/lib";
|
||||
# nativeBuildInputs = with pkgs; [
|
||||
# cmake
|
||||
# llvmPackages.libclang.lib
|
||||
# ];
|
||||
buildInputs = with pkgs;
|
||||
[]
|
||||
++ (lib.optionals pkgs.stdenv.isDarwin [
|
||||
libiconv
|
||||
apple-sdk_13
|
||||
]);
|
||||
}
|
||||
// (lib.optionalAttrs pkgs.stdenv.isLinux {
|
||||
# BINDGEN_EXTRA_CLANG_ARGS = "-I${pkgs.llvmPackages.libclang.lib}/lib/clang/18/include";
|
||||
});
|
||||
cargoArtifacts = craneLib.buildPackage commonArgs;
|
||||
in {
|
||||
checks =
|
||||
{
|
||||
"${name}-clippy" = craneLib.cargoClippy (commonArgs
|
||||
// {
|
||||
inherit cargoArtifacts;
|
||||
cargoClippyExtraArgs = "--all-targets -- --deny warnings";
|
||||
});
|
||||
"${name}-docs" = craneLib.cargoDoc (commonArgs // {inherit cargoArtifacts;});
|
||||
"${name}-fmt" = craneLib.cargoFmt {inherit src;};
|
||||
"${name}-toml-fmt" = craneLib.taploFmt {
|
||||
src = pkgs.lib.sources.sourceFilesBySuffices src [".toml"];
|
||||
};
|
||||
# Audit dependencies
|
||||
"${name}-audit" = craneLib.cargoAudit {
|
||||
inherit src advisory-db;
|
||||
};
|
||||
|
||||
# Audit licenses
|
||||
"${name}-deny" = craneLib.cargoDeny {
|
||||
inherit src;
|
||||
};
|
||||
"${name}-nextest" = craneLib.cargoNextest (commonArgs
|
||||
// {
|
||||
inherit cargoArtifacts;
|
||||
partitions = 1;
|
||||
partitionType = "count";
|
||||
});
|
||||
}
|
||||
// lib.optionalAttrs (!pkgs.stdenv.isDarwin) {
|
||||
"${name}-llvm-cov" = craneLibLLvmTools.cargoLlvmCov (commonArgs // {inherit cargoArtifacts;});
|
||||
};
|
||||
|
||||
packages = let
|
||||
pkg = craneLib.buildPackage (commonArgs
|
||||
// {inherit cargoArtifacts;}
|
||||
// {
|
||||
nativeBuildInputs = with pkgs; [
|
||||
installShellFiles
|
||||
];
|
||||
postInstall = ''
|
||||
installShellCompletion --cmd ${name} \
|
||||
--bash <($out/bin/${name} completions bash) \
|
||||
--fish <($out/bin/${name} completions fish) \
|
||||
--zsh <($out/bin/${name} completions zsh)
|
||||
'';
|
||||
});
|
||||
in {
|
||||
"${name}" = pkg;
|
||||
default = pkg;
|
||||
};
|
||||
|
||||
devShells = {
|
||||
default = pkgs.mkShell.override {stdenv = pkgs.clangStdenv;} (commonArgs
|
||||
// {
|
||||
packages = with pkgs;
|
||||
[
|
||||
stableToolchainWithRustAnalyzer
|
||||
cargo-nextest
|
||||
cargo-deny
|
||||
mnn
|
||||
]
|
||||
++ (lib.optionals pkgs.stdenv.isDarwin [
|
||||
apple-sdk_13
|
||||
]);
|
||||
});
|
||||
};
|
||||
}
|
||||
)
|
||||
// {
|
||||
githubActions = nix-github-actions.lib.mkGithubMatrix {
|
||||
checks = nixpkgs.lib.getAttrs ["x86_64-linux"] self.checks;
|
||||
};
|
||||
};
|
||||
}
|
||||
BIN
models/retinaface.mnn
Normal file
BIN
models/retinaface.mnn
Normal file
Binary file not shown.
1
ndarray-image/.gitignore
vendored
Normal file
1
ndarray-image/.gitignore
vendored
Normal file
@@ -0,0 +1 @@
|
||||
/target
|
||||
1132
ndarray-image/Cargo.lock
generated
Normal file
1132
ndarray-image/Cargo.lock
generated
Normal file
File diff suppressed because it is too large
Load Diff
8
ndarray-image/Cargo.toml
Normal file
8
ndarray-image/Cargo.toml
Normal file
@@ -0,0 +1,8 @@
|
||||
[package]
|
||||
name = "ndarray-image"
|
||||
version = "0.1.0"
|
||||
edition = "2024"
|
||||
|
||||
[dependencies]
|
||||
image = "0.25.6"
|
||||
ndarray = "0.16.1"
|
||||
181
ndarray-image/src/lib.rs
Normal file
181
ndarray-image/src/lib.rs
Normal file
@@ -0,0 +1,181 @@
|
||||
type Result<T, E = ndarray::ShapeError> = core::result::Result<T, E>;
|
||||
mod rgb8 {
|
||||
use super::Result;
|
||||
pub(super) fn image_as_ndarray(image: &image::RgbImage) -> Result<ndarray::ArrayView3<u8>> {
|
||||
let (width, height) = image.dimensions();
|
||||
let data = image.as_raw();
|
||||
ndarray::ArrayView3::from_shape((height as usize, width as usize, 3), data)
|
||||
}
|
||||
pub(super) fn image_into_ndarray(image: image::RgbImage) -> Result<ndarray::Array3<u8>> {
|
||||
let (width, height) = image.dimensions();
|
||||
let data = image.into_raw();
|
||||
ndarray::Array3::from_shape_vec((height as usize, width as usize, 3), data)
|
||||
}
|
||||
}
|
||||
|
||||
mod rgba8 {
|
||||
use super::Result;
|
||||
pub(super) fn image_as_ndarray(image: &image::RgbaImage) -> Result<ndarray::ArrayView3<u8>> {
|
||||
let (width, height) = image.dimensions();
|
||||
let data = image.as_raw();
|
||||
ndarray::ArrayView3::from_shape((height as usize, width as usize, 4), data)
|
||||
}
|
||||
pub(super) fn image_into_ndarray(image: image::RgbaImage) -> Result<ndarray::Array3<u8>> {
|
||||
let (width, height) = image.dimensions();
|
||||
let data = image.into_raw();
|
||||
ndarray::Array3::from_shape_vec((height as usize, width as usize, 4), data)
|
||||
}
|
||||
}
|
||||
|
||||
mod gray8 {
|
||||
use super::Result;
|
||||
pub(super) fn image_as_ndarray(image: &image::GrayImage) -> Result<ndarray::ArrayView2<u8>> {
|
||||
let (width, height) = image.dimensions();
|
||||
let data = image.as_raw();
|
||||
ndarray::ArrayView2::from_shape((height as usize, width as usize), data)
|
||||
}
|
||||
pub(super) fn image_into_ndarray(image: image::GrayImage) -> Result<ndarray::Array2<u8>> {
|
||||
let (width, height) = image.dimensions();
|
||||
let data = image.into_raw();
|
||||
ndarray::Array2::from_shape_vec((height as usize, width as usize), data)
|
||||
}
|
||||
}
|
||||
|
||||
mod gray_alpha8 {
|
||||
use super::Result;
|
||||
pub(super) fn image_as_ndarray(
|
||||
image: &image::GrayAlphaImage,
|
||||
) -> Result<ndarray::ArrayView3<u8>> {
|
||||
let (width, height) = image.dimensions();
|
||||
let data = image.as_raw();
|
||||
ndarray::ArrayView3::from_shape((height as usize, width as usize, 2), data)
|
||||
}
|
||||
pub(super) fn image_into_ndarray(image: image::GrayAlphaImage) -> Result<ndarray::Array3<u8>> {
|
||||
let (width, height) = image.dimensions();
|
||||
let data = image.into_raw();
|
||||
ndarray::Array3::from_shape_vec((height as usize, width as usize, 2), data)
|
||||
}
|
||||
}
|
||||
|
||||
mod dynamic_image {
|
||||
use super::*;
|
||||
pub fn image_as_ndarray(image: &image::DynamicImage) -> Result<ndarray::ArrayViewD<u8>> {
|
||||
Ok(match image {
|
||||
image::DynamicImage::ImageRgb8(img) => rgb8::image_as_ndarray(img)?.into_dyn(),
|
||||
image::DynamicImage::ImageRgba8(img) => rgba8::image_as_ndarray(img)?.into_dyn(),
|
||||
image::DynamicImage::ImageLuma8(img) => gray8::image_as_ndarray(img)?.into_dyn(),
|
||||
image::DynamicImage::ImageLumaA8(img) => gray_alpha8::image_as_ndarray(img)?.into_dyn(),
|
||||
_ => {
|
||||
unimplemented!("Unsupported image format: {:?}", image);
|
||||
}
|
||||
})
|
||||
}
|
||||
pub fn image_into_ndarray(image: image::DynamicImage) -> Result<ndarray::ArrayD<u8>> {
|
||||
Ok(match image {
|
||||
image::DynamicImage::ImageRgb8(img) => rgb8::image_into_ndarray(img)?.into_dyn(),
|
||||
image::DynamicImage::ImageRgba8(img) => rgba8::image_into_ndarray(img)?.into_dyn(),
|
||||
image::DynamicImage::ImageLuma8(img) => gray8::image_into_ndarray(img)?.into_dyn(),
|
||||
image::DynamicImage::ImageLumaA8(img) => {
|
||||
gray_alpha8::image_into_ndarray(img)?.into_dyn()
|
||||
}
|
||||
_ => {
|
||||
unimplemented!("Unsupported image format: {:?}", image);
|
||||
}
|
||||
})
|
||||
}
|
||||
}
|
||||
pub trait ImageToNdarray {
|
||||
type OwnedOutput;
|
||||
type RefOutput<'a>
|
||||
where
|
||||
Self: 'a;
|
||||
|
||||
fn as_ndarray<'a>(&'a self) -> Result<Self::RefOutput<'a>>;
|
||||
fn to_ndarray(&self) -> Result<Self::OwnedOutput>;
|
||||
fn into_ndarray(self) -> Result<Self::OwnedOutput>;
|
||||
}
|
||||
|
||||
impl ImageToNdarray for image::RgbImage {
|
||||
type OwnedOutput = ndarray::Array3<u8>;
|
||||
type RefOutput<'a> = ndarray::ArrayView3<'a, u8>;
|
||||
|
||||
fn as_ndarray<'a>(&'a self) -> Result<Self::RefOutput<'a>> {
|
||||
rgb8::image_as_ndarray(self)
|
||||
}
|
||||
|
||||
fn to_ndarray(&self) -> Result<Self::OwnedOutput> {
|
||||
Ok(self.as_ndarray()?.to_owned())
|
||||
}
|
||||
|
||||
fn into_ndarray(self) -> Result<Self::OwnedOutput> {
|
||||
rgb8::image_into_ndarray(self)
|
||||
}
|
||||
}
|
||||
|
||||
impl ImageToNdarray for image::RgbaImage {
|
||||
type OwnedOutput = ndarray::Array3<u8>;
|
||||
type RefOutput<'a> = ndarray::ArrayView3<'a, u8>;
|
||||
|
||||
fn as_ndarray<'a>(&'a self) -> Result<Self::RefOutput<'a>> {
|
||||
rgba8::image_as_ndarray(self)
|
||||
}
|
||||
|
||||
fn to_ndarray(&self) -> Result<Self::OwnedOutput> {
|
||||
Ok(self.as_ndarray()?.to_owned())
|
||||
}
|
||||
|
||||
fn into_ndarray(self) -> Result<Self::OwnedOutput> {
|
||||
rgba8::image_into_ndarray(self)
|
||||
}
|
||||
}
|
||||
|
||||
impl ImageToNdarray for image::GrayImage {
|
||||
type OwnedOutput = ndarray::Array2<u8>;
|
||||
type RefOutput<'a> = ndarray::ArrayView2<'a, u8>;
|
||||
|
||||
fn as_ndarray<'a>(&'a self) -> Result<Self::RefOutput<'a>> {
|
||||
gray8::image_as_ndarray(self)
|
||||
}
|
||||
|
||||
fn to_ndarray(&self) -> Result<Self::OwnedOutput> {
|
||||
Ok(self.as_ndarray()?.to_owned())
|
||||
}
|
||||
|
||||
fn into_ndarray(self) -> Result<Self::OwnedOutput> {
|
||||
gray8::image_into_ndarray(self)
|
||||
}
|
||||
}
|
||||
|
||||
impl ImageToNdarray for image::GrayAlphaImage {
|
||||
type OwnedOutput = ndarray::Array3<u8>;
|
||||
type RefOutput<'a> = ndarray::ArrayView3<'a, u8>;
|
||||
|
||||
fn as_ndarray<'a>(&'a self) -> Result<Self::RefOutput<'a>> {
|
||||
gray_alpha8::image_as_ndarray(self)
|
||||
}
|
||||
|
||||
fn to_ndarray(&self) -> Result<Self::OwnedOutput> {
|
||||
Ok(self.as_ndarray()?.to_owned())
|
||||
}
|
||||
|
||||
fn into_ndarray(self) -> Result<Self::OwnedOutput> {
|
||||
gray_alpha8::image_into_ndarray(self)
|
||||
}
|
||||
}
|
||||
|
||||
impl ImageToNdarray for image::DynamicImage {
|
||||
type OwnedOutput = ndarray::ArrayD<u8>;
|
||||
type RefOutput<'a> = ndarray::ArrayViewD<'a, u8>;
|
||||
|
||||
fn as_ndarray<'a>(&'a self) -> Result<Self::RefOutput<'a>> {
|
||||
dynamic_image::image_as_ndarray(self)
|
||||
}
|
||||
|
||||
fn to_ndarray(&self) -> Result<Self::OwnedOutput> {
|
||||
Ok(self.as_ndarray()?.to_owned())
|
||||
}
|
||||
|
||||
fn into_ndarray(self) -> Result<Self::OwnedOutput> {
|
||||
dynamic_image::image_into_ndarray(self)
|
||||
}
|
||||
}
|
||||
64
src/cli.rs
Normal file
64
src/cli.rs
Normal file
@@ -0,0 +1,64 @@
|
||||
use std::path::PathBuf;
|
||||
#[derive(Debug, clap::Parser)]
|
||||
pub struct Cli {
|
||||
#[clap(subcommand)]
|
||||
pub cmd: SubCommand,
|
||||
}
|
||||
|
||||
#[derive(Debug, clap::Subcommand)]
|
||||
pub enum SubCommand {
|
||||
#[clap(name = "detect")]
|
||||
Detect(Detect),
|
||||
#[clap(name = "list")]
|
||||
List(List),
|
||||
#[clap(name = "completions")]
|
||||
Completions { shell: clap_complete::Shell },
|
||||
}
|
||||
|
||||
#[derive(Debug, clap::ValueEnum, Clone, Copy)]
|
||||
pub enum Models {
|
||||
RetinaFace,
|
||||
}
|
||||
|
||||
#[derive(Debug, clap::ValueEnum, Clone, Copy)]
|
||||
pub enum Executor {
|
||||
Mnn,
|
||||
Onnx,
|
||||
}
|
||||
|
||||
#[derive(Debug, clap::ValueEnum, Clone, Copy)]
|
||||
pub enum OnnxEp {
|
||||
Cpu,
|
||||
}
|
||||
|
||||
#[derive(Debug, clap::ValueEnum, Clone, Copy)]
|
||||
pub enum MnnEp {
|
||||
Cpu,
|
||||
Metal,
|
||||
OpenCL,
|
||||
CoreML,
|
||||
}
|
||||
|
||||
#[derive(Debug, clap::Args)]
|
||||
pub struct Detect {
|
||||
#[clap(short, long)]
|
||||
pub model: Option<PathBuf>,
|
||||
#[clap(short = 'M', long, default_value = "retina-face")]
|
||||
pub model_type: Models,
|
||||
pub image: PathBuf,
|
||||
}
|
||||
|
||||
#[derive(Debug, clap::Args)]
|
||||
pub struct List {}
|
||||
|
||||
impl Cli {
|
||||
pub fn completions(shell: clap_complete::Shell) {
|
||||
let mut command = <Cli as clap::CommandFactory>::command();
|
||||
clap_complete::generate(
|
||||
shell,
|
||||
&mut command,
|
||||
env!("CARGO_BIN_NAME"),
|
||||
&mut std::io::stdout(),
|
||||
);
|
||||
}
|
||||
}
|
||||
6
src/errors.rs
Normal file
6
src/errors.rs
Normal file
@@ -0,0 +1,6 @@
|
||||
pub use error_stack::{Report, ResultExt};
|
||||
#[derive(Debug, thiserror::Error)]
|
||||
#[error("An error occurred")]
|
||||
pub struct Error;
|
||||
|
||||
pub type Result<T, E = error_stack::Report<Error>> = core::result::Result<T, E>;
|
||||
73
src/facedet.rs
Normal file
73
src/facedet.rs
Normal file
@@ -0,0 +1,73 @@
|
||||
use crate::errors::*;
|
||||
use error_stack::ResultExt;
|
||||
use mnn_bridge::ndarray::NdarrayToMnn;
|
||||
use std::path::Path;
|
||||
|
||||
pub struct FaceDetection {
|
||||
handle: mnn_sync::SessionHandle,
|
||||
}
|
||||
|
||||
impl FaceDetection {
|
||||
pub fn new(path: impl AsRef<Path>) -> Result<Self> {
|
||||
let model = std::fs::read(path)
|
||||
.change_context(Error)
|
||||
.attach_printable("Failed to read model file")?;
|
||||
Self::new_from_bytes(&model)
|
||||
}
|
||||
|
||||
pub fn new_from_bytes(model: &[u8]) -> Result<Self> {
|
||||
tracing::info!("Loading face detection model from bytes");
|
||||
let mut model = mnn::Interpreter::from_bytes(model)
|
||||
.map_err(|e| e.into_inner())
|
||||
.change_context(Error)
|
||||
.attach_printable("Failed to load model from bytes")?;
|
||||
model.set_session_mode(mnn::SessionMode::Release);
|
||||
let bc = mnn::BackendConfig::default().with_memory_mode(mnn::MemoryMode::High);
|
||||
let sc = mnn::ScheduleConfig::new()
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
.with_type(mnn::ForwardType::CPU)
|
||||
.with_backend_config(bc);
|
||||
tracing::info!("Creating session handle for face detection model");
|
||||
let handle = mnn_sync::SessionHandle::new(model, sc)
|
||||
.change_context(Error)
|
||||
.attach_printable("Failed to create session handle")?;
|
||||
Ok(FaceDetection { handle })
|
||||
}
|
||||
|
||||
pub fn detect_faces(&self, image: ndarray::Array3<u8>) -> Result<ndarray::Array2<u8>> {
|
||||
use mnn_bridge::ndarray::MnnToNdarray;
|
||||
let output = self
|
||||
.handle
|
||||
.run(move |sr| {
|
||||
let tensor = image
|
||||
.as_mnn_tensor()
|
||||
.ok_or_else(|| Error)
|
||||
.attach_printable("Failed to convert ndarray to mnn tensor")
|
||||
.change_context(mnn::error::ErrorKind::TensorError)?;
|
||||
let (intptr, session) = sr.both_mut();
|
||||
tracing::trace!("Copying input tensor to host");
|
||||
// let input = intptr.input::<u8>(session, "input")?;
|
||||
// dbg!(input.shape());
|
||||
// let mut t = input.create_host_tensor_from_device(false);
|
||||
// tensor.copy_to_host_tensor(&mut t)?;
|
||||
//
|
||||
// intptr.run_session(&session)?;
|
||||
// let output = intptr.output::<u8>(&session, "output").unwrap();
|
||||
// let output_tensor = output.create_host_tensor_from_device(true);
|
||||
// let output_array = output_tensor
|
||||
// .try_as_ndarray()
|
||||
// .change_context(mnn::error::ErrorKind::TensorError)?
|
||||
// .to_owned();
|
||||
// Ok(output_array)
|
||||
Ok(ndarray::Array2::<u8>::zeros((1, 1))) // Placeholder for actual output
|
||||
})
|
||||
.map_err(|e| e.into_inner())
|
||||
.change_context(Error);
|
||||
output
|
||||
}
|
||||
}
|
||||
5
src/image.rs
Normal file
5
src/image.rs
Normal file
@@ -0,0 +1,5 @@
|
||||
// pub struct Image {
|
||||
// pub width: u32,
|
||||
// pub height: u32,
|
||||
// pub data: Vec<u8>,
|
||||
// }
|
||||
4
src/lib.rs
Normal file
4
src/lib.rs
Normal file
@@ -0,0 +1,4 @@
|
||||
pub mod errors;
|
||||
pub mod facedet;
|
||||
pub mod image;
|
||||
use errors::*;
|
||||
37
src/main.rs
Normal file
37
src/main.rs
Normal file
@@ -0,0 +1,37 @@
|
||||
mod cli;
|
||||
mod errors;
|
||||
use errors::*;
|
||||
use ndarray_image::*;
|
||||
const RETINAFACE_MODEL: &[u8] = include_bytes!("../models/retinaface.mnn");
|
||||
pub fn main() -> Result<()> {
|
||||
tracing_subscriber::fmt()
|
||||
.with_env_filter("trace")
|
||||
.with_thread_ids(true)
|
||||
.with_thread_names(true)
|
||||
.with_target(false)
|
||||
.init();
|
||||
let args = <cli::Cli as clap::Parser>::parse();
|
||||
match args.cmd {
|
||||
cli::SubCommand::Detect(detect) => {
|
||||
use detector::facedet;
|
||||
let model = facedet::FaceDetection::new_from_bytes(RETINAFACE_MODEL)
|
||||
.change_context(errors::Error)
|
||||
.attach_printable("Failed to create face detection model")?;
|
||||
let image = image::open(detect.image).change_context(Error)?;
|
||||
let image = image.into_rgb8();
|
||||
let array = image.into_ndarray()
|
||||
.change_context(errors::Error)
|
||||
.attach_printable("Failed to convert image to ndarray")?;
|
||||
model.detect_faces(array)
|
||||
.change_context(errors::Error)
|
||||
.attach_printable("Failed to detect faces")?;
|
||||
}
|
||||
cli::SubCommand::List(list) => {
|
||||
println!("List: {:?}", list);
|
||||
}
|
||||
cli::SubCommand::Completions { shell } => {
|
||||
cli::Cli::completions(shell);
|
||||
}
|
||||
}
|
||||
Ok(())
|
||||
}
|
||||
Reference in New Issue
Block a user