1
Fork 0

Migrate more stuff towards my own ocr

Signed-off-by: prescientmoon <git@moonythm.dev>
This commit is contained in:
prescientmoon 2024-08-11 18:32:17 +02:00
parent 8ffe75a23f
commit 5c95cdb018
Signed by: prescientmoon
SSH key fingerprint: SHA256:UUF9JT2s8Xfyv76b8ZuVL7XrmimH4o49p4b+iexbVH4
6 changed files with 176 additions and 125 deletions

View file

@ -1,9 +1,12 @@
use std::time::Instant;
use crate::arcaea::play::{CreatePlay, Play};
use crate::arcaea::score::Score;
use crate::context::{Context, Error};
use crate::recognition::recognize::{ImageAnalyzer, ScoreKind};
use crate::user::{discord_it_to_discord_user, User};
use crate::{edit_reply, get_user};
use crate::{edit_reply, get_user, timed};
use image::DynamicImage;
use poise::serenity_prelude::CreateMessage;
use poise::{serenity_prelude as serenity, CreateReply};
use sqlx::query;
@ -41,52 +44,61 @@ pub async fn magic(
let mut analyzer = ImageAnalyzer::default();
for (i, file) in files.iter().enumerate() {
let start = Instant::now();
if let Some(_) = file.dimensions() {
let bytes = file.download().await?;
let mut image = image::load_from_memory(&bytes)?;
let bytes = timed!("file download", { file.download().await? });
let mut image = timed!("decode image", { image::load_from_memory(&bytes)? });
let mut grayscale_image = timed!("grayscale image", {
DynamicImage::ImageLuma8(image.to_luma8())
});
// image = image.resize(1024, 1024, FilterType::Nearest);
let result: Result<(), Error> = try {
// {{{ Detection
// This makes OCR more likely to work
let mut ocr_image = image.grayscale().blur(1.);
edit_reply!(ctx, handle, "Image {}: reading kind", i + 1).await?;
let kind = analyzer.read_score_kind(ctx.data(), &ocr_image)?;
// edit_reply!(ctx, handle, "Image {}: reading kind", i + 1).await?;
let kind = timed!("read_score_kind", {
analyzer.read_score_kind(ctx.data(), &grayscale_image)?
});
edit_reply!(ctx, handle, "Image {}: reading difficulty", i + 1).await?;
// edit_reply!(ctx, handle, "Image {}: reading difficulty", i + 1).await?;
// Do not use `ocr_image` because this reads the colors
let difficulty = analyzer.read_difficulty(ctx.data(), &image, kind)?;
let difficulty = timed!("read_difficulty", {
analyzer.read_difficulty(ctx.data(), &image, kind)?
});
edit_reply!(ctx, handle, "Image {}: reading jacket", i + 1).await?;
let (song, chart) = analyzer
.read_jacket(ctx.data(), &mut image, kind, difficulty)
.await?;
ocr_image.invert();
// edit_reply!(ctx, handle, "Image {}: reading jacket", i + 1).await?;
let (song, chart) = timed!("read_jacket", {
analyzer.read_jacket(ctx.data(), &mut image, kind, difficulty)?
});
let (note_distribution, max_recall) = match kind {
ScoreKind::ScoreScreen => {
edit_reply!(ctx, handle, "Image {}: reading distribution", i + 1)
.await?;
let note_distribution =
Some(analyzer.read_distribution(ctx.data(), &image)?);
Some(analyzer.read_distribution(ctx.data(), &grayscale_image)?);
edit_reply!(ctx, handle, "Image {}: reading max recall", i + 1).await?;
let max_recall = Some(analyzer.read_max_recall(ctx.data(), &image)?);
let max_recall =
Some(analyzer.read_max_recall(ctx.data(), &grayscale_image)?);
(note_distribution, max_recall)
}
ScoreKind::SongSelect => (None, None),
};
edit_reply!(ctx, handle, "Image {}: reading score", i + 1).await?;
let score = analyzer.read_score(
grayscale_image.invert();
// edit_reply!(ctx, handle, "Image {}: reading score", i + 1).await?;
let score = timed!("read_score", {
analyzer.read_score(
ctx.data(),
Some(chart.note_count),
&ocr_image,
&grayscale_image,
kind,
)?;
)?
});
// {{{ Build play
let maybe_fars = Score::resolve_distibution_ambiguities(
@ -104,9 +116,11 @@ pub async fn magic(
// }}}
// }}}
// {{{ Deliver embed
let (embed, attachment) = play
.to_embed(&ctx.data().db, &user, &song, &chart, i, None)
.await?;
let (embed, attachment) = timed!("to embed", {
play.to_embed(&ctx.data().db, &user, &song, &chart, i, None)
.await?
});
embeds.push(embed);
attachments.extend(attachment);
@ -123,8 +137,16 @@ pub async fn magic(
.await?;
continue;
}
let took = start.elapsed();
edit_reply!(ctx, handle, "Processed {}/{} scores", i + 1, files.len()).await?;
edit_reply!(
ctx,
handle,
"Processed {}/{} scores. Last score took {took:?} to process.",
i + 1,
files.len()
)
.await?;
}
handle.delete(ctx).await?;

View file

@ -35,10 +35,12 @@ impl UserContext {
let mut song_cache = SongCache::new(&db).await?;
let jacket_cache = JacketCache::new(&data_dir, &mut song_cache)?;
let ui_measurements = UIMeasurements::read(&data_dir)?;
let geosans_measurements = GEOSANS_FONT
.with_borrow_mut(|font| CharMeasurements::from_text(font, "0123456789'", None))?;
let exo_measurements = EXO_FONT
.with_borrow_mut(|font| CharMeasurements::from_text(font, "0123456789'", Some(700)))?;
let exo_measurements = EXO_FONT.with_borrow_mut(|font| {
CharMeasurements::from_text(font, "0123456789'abcdefghijklmnopqrstuvwxyz", Some(700))
})?;
println!("Created user context");

View file

@ -13,6 +13,7 @@ mod context;
mod levenshtein;
mod logs;
mod recognition;
mod time;
mod transform;
mod user;

View file

@ -12,7 +12,7 @@
//! 5. Compute the largest width & height of the connected components.
//! 5. Split each component (more precisely, start at its top-left corner and
//! split an area equal to the aforementioned width & height) into a grid of
//! N^2 chunks (N=5 at the moment), and use that to generate a vector who's
//! N^2 chunks (N=5 at the moment), and use that to generate a vector whose
//! elements represent the percentage of black pixels in each chunk which
//! belong to the connected component at hand.
//! 6. Normalise the vectors to remain font-weight independent.
@ -33,6 +33,7 @@ use crate::{
bitmap::{Align, BitmapCanvas, Color, TextStyle},
context::Error,
logs::{debug_image_buffer_log, debug_image_log},
timed,
};
// {{{ ConponentVec
@ -47,7 +48,11 @@ struct ComponentVec {
impl ComponentVec {
// {{{ (Component => vector) encoding
fn from_component(components: &ComponentsWithBounds, component: u32) -> Result<Self, Error> {
fn from_component(
components: &ComponentsWithBounds,
area: (u32, u32),
component: u32,
) -> Result<Self, Error> {
let mut chunks = [0.0; IMAGE_VEC_DIM];
let bounds = components
.bounds
@ -58,10 +63,10 @@ impl ComponentVec {
for i in 0..(SPLIT_FACTOR * SPLIT_FACTOR) {
let (iy, ix) = i.div_rem_euclid(&SPLIT_FACTOR);
let x_start = bounds.x_min + ix * components.max_width / SPLIT_FACTOR;
let x_end = bounds.x_min + (ix + 1) * components.max_width / SPLIT_FACTOR;
let y_start = bounds.y_min + iy * components.max_height / SPLIT_FACTOR;
let y_end = bounds.y_min + (iy + 1) * components.max_height / SPLIT_FACTOR;
let x_start = bounds.x_min + ix * area.0 / SPLIT_FACTOR;
let x_end = bounds.x_min + (ix + 1) * area.0 / SPLIT_FACTOR;
let y_start = bounds.y_min + iy * area.1 / SPLIT_FACTOR;
let y_end = bounds.y_min + (iy + 1) * area.1 / SPLIT_FACTOR;
let mut count = 0;
for x in x_start..x_end {
@ -148,9 +153,6 @@ struct ComponentsWithBounds {
// but we don't want to waste a place in this vector.
bounds: Vec<Option<ComponentBounds>>,
max_width: u32,
max_height: u32,
/// Stores the indices of `self.bounds` sorted based on their min position.
bounds_by_position: Vec<usize>,
}
@ -202,20 +204,6 @@ impl ComponentsWithBounds {
}
}
// }}}
// {{{ Compute max width/height
let max_width = bounds
.iter()
.filter_map(|o| o.as_ref())
.map(|b| b.x_max - b.x_min)
.max()
.ok_or_else(|| "No connected components found")?;
let max_height = bounds
.iter()
.filter_map(|o| o.as_ref())
.map(|b| b.y_max - b.y_min)
.max()
.ok_or_else(|| "No connected components found")?;
// }}}
let mut bounds_by_position: Vec<usize> = (0..(bounds.len()))
.filter(|i| bounds[*i].is_some())
@ -225,8 +213,6 @@ impl ComponentsWithBounds {
Ok(Self {
components,
bounds,
max_width,
max_height,
bounds_by_position,
})
}
@ -235,11 +221,14 @@ impl ComponentsWithBounds {
// {{{ Char measurements
pub struct CharMeasurements {
chars: Vec<(char, ComponentVec)>,
max_width: u32,
max_height: u32,
}
impl CharMeasurements {
// {{{ Creation
pub fn from_text(face: &mut Face, string: &str, _weight: Option<u32>) -> Result<Self, Error> {
pub fn from_text(face: &mut Face, string: &str, weight: Option<u32>) -> Result<Self, Error> {
// These are bad estimates lol
let char_w = 35;
let char_h = 60;
@ -255,7 +244,7 @@ impl CharMeasurements {
size: char_h,
color: Color::BLACK,
// TODO: do we want to use the weight hint for resilience?
weight: None,
weight,
},
&string,
)?;
@ -267,30 +256,64 @@ impl CharMeasurements {
let components = ComponentsWithBounds::from_image(&image)?;
// {{{ Compute max width/height
let max_width = components
.bounds
.iter()
.filter_map(|o| o.as_ref())
.map(|b| b.x_max - b.x_min)
.max()
.ok_or_else(|| "No connected components found")?;
let max_height = components
.bounds
.iter()
.filter_map(|o| o.as_ref())
.map(|b| b.y_max - b.y_min)
.max()
.ok_or_else(|| "No connected components found")?;
// }}}
let mut chars = Vec::with_capacity(string.len());
for (i, char) in string.chars().enumerate() {
chars.push((
char,
ComponentVec::from_component(
&components,
(max_width, max_height),
components.bounds_by_position[i] as u32 + 1,
)?,
))
}
Ok(Self { chars })
Ok(Self {
chars,
max_width,
max_height,
})
}
// }}}
// {{{ Recognition
pub fn recognise(&self, image: &DynamicImage) -> Result<String, Error> {
let components = ComponentsWithBounds::from_image(image)?;
pub fn recognise(&self, image: &DynamicImage, whitelist: &str) -> Result<String, Error> {
let components = timed!("from_image", { ComponentsWithBounds::from_image(image)? });
let mut result = String::new();
let max_height = components
.bounds
.iter()
.filter_map(|o| o.as_ref())
.map(|b| b.y_max - b.y_min)
.max()
.ok_or_else(|| "No connected components found")?;
let max_width = self.max_width * max_height / self.max_height;
for i in &components.bounds_by_position {
let vec = ComponentVec::from_component(&components, *i as u32 + 1)?;
let vec =
ComponentVec::from_component(&components, (max_width, max_height), *i as u32 + 1)?;
let best_match = self
.chars
.iter()
.filter(|(c, _)| whitelist.contains(*c))
.map(|(i, v)| (*i, v, v.distance_squared_to(&vec)))
.min_by(|(_, _, d1), (_, _, d2)| {
d1.partial_cmp(d2).expect("NaN distance encountered")

View file

@ -4,7 +4,6 @@ use std::str::FromStr;
use hypertesseract::{PageSegMode, Tesseract};
use image::imageops::FilterType;
use image::{DynamicImage, GenericImageView};
use image::{ImageBuffer, Rgba};
use num::integer::Roots;
use poise::serenity_prelude::{CreateAttachment, CreateEmbed, CreateMessage};
@ -14,11 +13,12 @@ use crate::arcaea::score::Score;
use crate::bitmap::{Color, Rect};
use crate::context::{Context, Error, UserContext};
use crate::levenshtein::edit_distance;
use crate::logs::debug_image_buffer_log;
use crate::logs::debug_image_log;
use crate::recognition::fuzzy_song_name::guess_chart_name;
use crate::recognition::ui::{
ScoreScreenRect, SongSelectRect, UIMeasurementRect, UIMeasurementRect::*,
};
use crate::timed;
use crate::transform::rotate;
#[derive(Debug, Clone, Copy, PartialEq, Eq)]
@ -47,10 +47,8 @@ impl ImageAnalyzer {
// {{{ Crop
#[inline]
pub fn crop(&mut self, image: &DynamicImage, rect: Rect) -> ImageBuffer<Rgba<u8>, Vec<u8>> {
image
.crop_imm(rect.x as u32, rect.y as u32, rect.width, rect.height)
.to_rgba8()
pub fn crop(&mut self, image: &DynamicImage, rect: Rect) -> DynamicImage {
image.crop_imm(rect.x as u32, rect.y as u32, rect.width, rect.height)
}
#[inline]
@ -59,12 +57,12 @@ impl ImageAnalyzer {
ctx: &UserContext,
image: &DynamicImage,
ui_rect: UIMeasurementRect,
) -> Result<ImageBuffer<Rgba<u8>, Vec<u8>>, Error> {
) -> Result<DynamicImage, Error> {
let rect = ctx.ui_measurements.interpolate(ui_rect, image)?;
self.last_rect = Some((ui_rect, rect));
let result = self.crop(image, rect);
debug_image_buffer_log(&result)?;
debug_image_log(&result)?;
Ok(result)
}
@ -76,16 +74,14 @@ impl ImageAnalyzer {
image: &DynamicImage,
ui_rect: UIMeasurementRect,
size: (u32, u32),
) -> Result<ImageBuffer<Rgba<u8>, Vec<u8>>, Error> {
) -> Result<DynamicImage, Error> {
let rect = ctx.ui_measurements.interpolate(ui_rect, image)?;
self.last_rect = Some((ui_rect, rect));
let result = self.crop(image, rect);
let result = DynamicImage::ImageRgba8(result)
.resize(size.0, size.1, FilterType::Nearest)
.into_rgba8();
let result = result.resize(size.0, size.1, FilterType::Nearest);
debug_image_buffer_log(&result)?;
debug_image_log(&result)?;
Ok(result)
}
@ -138,7 +134,8 @@ impl ImageAnalyzer {
image: &DynamicImage,
kind: ScoreKind,
) -> Result<Score, Error> {
let image = self.interp_crop_resize(
let image = timed!("interp_crop_resize", {
self.interp_crop_resize(
ctx,
image,
match kind {
@ -146,24 +143,26 @@ impl ImageAnalyzer {
ScoreKind::ScoreScreen => ScoreScreen(ScoreScreenRect::Score),
},
(u32::MAX, 100),
)?;
)?
});
let measurements = match kind {
ScoreKind::SongSelect => &ctx.exo_measurements,
ScoreKind::ScoreScreen => &ctx.geosans_measurements,
};
let result = Score(
let result = timed!("full recognition", {
Score(
measurements
.recognise(&DynamicImage::ImageRgba8(image))?
.recognise(&image, "0123456789'")?
.chars()
.filter(|c| *c != '\'')
.collect::<String>()
.parse()?,
);
)
});
// {{{ Return score if consensus exists
// 1. Discard scores that are known to be impossible
// Discard scores if it's impossible
if result.0 <= 10_010_000
&& note_count.map_or(true, |note_count| {
let (zeta, shinies, score_units) = result.analyse(note_count);
@ -222,11 +221,11 @@ impl ImageAnalyzer {
.language(hypertesseract::Language::English)
.page_seg_mode(PageSegMode::RawLine)
.build()?
.recognize_text_cloned_with_conf(&self.interp_crop(
ctx,
image,
ScoreScreen(ScoreScreenRect::Difficulty),
)?)?;
.recognize_text_cloned_with_conf(
&self
.interp_crop(ctx, image, ScoreScreen(ScoreScreenRect::Difficulty))?
.into_rgba8(),
)?;
let text = text.trim().to_lowercase();
@ -254,21 +253,10 @@ impl ImageAnalyzer {
ctx: &UserContext,
image: &DynamicImage,
) -> Result<ScoreKind, Error> {
let (text, conf) = Tesseract::builder()
.language(hypertesseract::Language::English)
.page_seg_mode(PageSegMode::RawLine)
.build()?
.recognize_text_cloned_with_conf(&self.interp_crop(ctx, image, PlayKind)?)?;
let text = text.trim().to_string();
if conf < 10 && conf != 0 {
return Err(format!(
"Score kind text is not readable (confidence = {}, text = {}).",
conf, text
)
.into());
}
let image = self.interp_crop(ctx, image, PlayKind)?;
let text = ctx
.exo_measurements
.recognise(&image, "resultselectasong")?;
let result = if edit_distance(&text, "Result") < edit_distance(&text, "Select a song") {
ScoreKind::ScoreScreen
@ -291,11 +279,11 @@ impl ImageAnalyzer {
.page_seg_mode(PageSegMode::SingleLine)
.whitelist_str("abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ0123456789,.()- ")?
.build()?
.recognize_text_cloned_with_conf(&self.interp_crop(
ctx,
image,
ScoreScreen(ScoreScreenRect::Title),
)?)?;
.recognize_text_cloned_with_conf(
&self
.interp_crop(ctx, image, ScoreScreen(ScoreScreenRect::Title))?
.into_rgba8(),
)?;
if conf < 20 && conf != 0 {
return Err(format!(
@ -310,7 +298,7 @@ impl ImageAnalyzer {
}
// }}}
// {{{ Read jacket
pub async fn read_jacket<'a>(
pub fn read_jacket<'a>(
&mut self,
ctx: &'a UserContext,
image: &mut DynamicImage,
@ -374,7 +362,11 @@ impl ImageAnalyzer {
.whitelist_str("0123456789")?
.assume_numeric_input()
.build()?
.recognize_text_cloned(&self.interp_crop(ctx, image, ScoreScreen(KINDS[i]))?)?;
.recognize_text_cloned(
&self
.interp_crop(ctx, image, ScoreScreen(KINDS[i]))?
.into_rgba8(),
)?;
println!("Raw '{}'", text.trim());
out[i] = u32::from_str(&text.trim()).unwrap_or(0);
@ -396,11 +388,11 @@ impl ImageAnalyzer {
.whitelist_str("0123456789")?
.assume_numeric_input()
.build()?
.recognize_text_cloned_with_conf(&self.interp_crop(
ctx,
image,
ScoreScreen(ScoreScreenRect::MaxRecall),
)?)?;
.recognize_text_cloned_with_conf(
&self
.interp_crop(ctx, image, ScoreScreen(ScoreScreenRect::MaxRecall))?
.into_rgba8(),
)?;
let max_recall = u32::from_str_radix(text.trim(), 10)?;

11
src/time.rs Normal file
View file

@ -0,0 +1,11 @@
#[macro_export]
macro_rules! timed {
($label:expr, $code:block) => {{
use std::time::Instant;
let start = Instant::now();
let result = { $code }; // Execute the code block
let duration = start.elapsed();
println!("{}: {:?}", $label, duration);
result
}};
}