- 02 Apr, 2025 1 commit
-
-
myhloli authored
- Remove unused imports for concurrent.futures, multiprocessing, and paddle - Delete commented-out code - Update numpy dependency to remove upper version limit - Remove InferenceResult import that was commented out
-
- 31 Mar, 2025 1 commit
-
-
myhloli authored
- Split OCR process into detection and recognition stages - Update batch analysis and document analysis pipelines - Modify OCR result formatting and handling - Remove unused imports and optimize code structure
-
- 26 Mar, 2025 2 commits
- 24 Mar, 2025 1 commit
-
-
icecraft authored
-
- 20 Mar, 2025 4 commits
-
-
myhloli authored
- Remove separate condition for GPU memory >= 24GB - Simplify logic to use a single threshold of 16GB
-
myhloli authored
- Increase batch ratio to 32 for GPU memory >= 24GB - Set batch ratio to 16 for GPU memory >= 16GB - Reduce batch ratio to 8 for GPU memory >= 12GB - Lower batch ratio to 4 for GPU memory >= 8GB - Set batch ratio to 2 for GPU memory >= 6GB - Keep batch ratio at 1 for lower GPU memory sizes
-
myhloli authored
- Remove torchtext version check and deprecation warning handling from multiple files - This code was unnecessary and potentially caused issues when torchtext was not installed
-
myhloli authored
- Add support for Apple M1 chips (mps device) - Refactor image processing for better performance and compatibility - Update model loading and inference for various devices - Adjust batch processing and memory management
-
- 19 Mar, 2025 1 commit
-
-
icecraft authored
-
- 13 Mar, 2025 2 commits
- 11 Mar, 2025 1 commit
-
-
myhloli authored
- Set NPUDTCompile to false for better performance on NPU - Adjust batch ratio
-
- 03 Mar, 2025 3 commits
-
-
myhloli authored
- Increase batch ratio to 8 for GPU memory >=16GB - Improve inference performance on systems with higher GPU memory
-
myhloli authored
- Simplify batch ratio logic for GPU memory >= 16GB - Remove unnecessary conditions for 20GB and 40GB memory
-
myhloli authored
- Simplify batch ratio logic for GPU memory >= 16GB - Remove unnecessary conditions for 20GB and 40GB memory
-
- 25 Feb, 2025 1 commit
-
-
myhloli authored
- Move batch model initialization outside the loop - Collect page dimensions before analyzing- Update page info dictionary structure - Add null dimensions for non-analyzed pages
-
- 24 Feb, 2025 1 commit
-
-
myhloli authored
- Update the logic for determining `end_page_id` to handle negative values - This change ensures proper behavior when `end_page_id` is set to -1 or other negative values
-
- 22 Feb, 2025 1 commit
-
-
Nathan Dahlberg authored
-
- 11 Feb, 2025 2 commits
-
-
myhloli authored
- Move environment variable settings for NPU, MPS, and other configurations to the global scope in doc_analyze_by_custom_model.py - Remove redundant environment variable settings in pdf_extract_kit.py - This change ensures consistent configuration across the application and avoids potential conflicts or duplicate settings
-
myhloli authored
-
- 10 Feb, 2025 1 commit
-
-
myhloli authored
- Remove unused utility functions - Update import statements for better readability - Add conditional imports for Ascend plugin - Refactor table model initialization to support NPU
-
- 07 Feb, 2025 1 commit
-
-
myhloli authored
- Update batch ratio calculation logic to better utilize available GPU memory - Improve logging for all GPU memory sizes
-
- 27 Jan, 2025 2 commits
- 21 Jan, 2025 6 commits
-
-
myhloli authored
- Update conditions for batch ratio assignment: -8 <= gpu_memory < 10: batch_ratio = 2 - 10 <= gpu_memory <= 12: batch_ratio =4 - This fix ensures proper batch ratio selection for GPU memory sizes
-
myhloli authored
- Improve batch ratio calculation based on GPU memory - Enhance performance for devices with 8GB or more VRAM
-
myhloli authored
- Reduce batch_ratio by 1 for better performance and stability - This change ensures more consistent memory usage when processing documents
-
myhloli authored
refactor(magic_pdf): adjust VRAM allocation and MFR batch size- Update VRAM allocation logic to use 'VIRTUAL_VRAM_SIZE' environment variable - Reduce MFR (Math Formula Recognition) batch size from 64 to 32
-
myhloli authored
- Update GPU memory check and batch ratio calculation logic - Add support for virtual VRAM size environment variable - Improve logging for GPU memory and batch ratio
-
myhloli authored
- Reduce YOLO_LAYOUT_BASE_BATCH_SIZE from 4 to 1 - Simplify batch ratio calculation for formula detection - Remove unused conditional logic in batch ratio determination
-
- 17 Jan, 2025 1 commit
-
-
myhloli authored
- Commented out the original batch ratio calculation - Set a fixed batch ratio of 2 for GPUs with less than 8 GB memory - Increased batch ratio to 4 for GPUs with 8 GB or more memory
-
- 16 Jan, 2025 1 commit
-
-
myhloli authored
- Adjust end_page_id calculation to prevent IndexError when accessing pages - Enhance error handling in LLM post-processing by specifically catching JSONDecodeError
-
- 15 Jan, 2025 1 commit
-
-
myhloli authored
- Add support for NPU (Neural Processing Unit) when available - Implement batch analysis for GPU and NPU devices - Optimize memory usage and improve performance - Update logging and error handling
-
- 26 Dec, 2024 1 commit
-
-
myhloli authored
- Update clean_memory function to support both CUDA and NPU devices - Implement get_device function to centralize device selection logic - Modify model initialization and memory cleaning to use the selected device - Update RapidTableModel to support both RapidOCR and PaddleOCR engines
-
- 18 Dec, 2024 1 commit
-
-
icecraft authored
-
- 17 Dec, 2024 1 commit
-
-
myhloli authored
- Add YOLOv11 language detection model for PDF documents - Implement language detection in PymuDocDataset - Update app.py to include 'auto' language option - Create language detection utilities and constants
-
- 16 Dec, 2024 1 commit
-
-
myhloli authored
- Remove YOLO_VERBOSE environment variable from multiple files - Set verbose=False in YOLOv8 prediction method to suppress logger output
-
- 10 Dec, 2024 1 commit
-
-
myhloli authored
- Import paddle module and disable its signal handler to prevent interference with other components - This change addresses potential conflicts between PaddlePaddle and other libraries or system signals
-
- 09 Dec, 2024 1 commit
-
-
myhloli authored
- Add environment variables to disable albumentations and yolo updates - Import torchtext and disable deprecation warnings - Update unimernet to 0.2.2 - Specify ultralytics version as >=8.3.48 - Remove upper version limit for torch
-