mirror of
https://github.com/we-promise/sure.git
synced 2026-04-09 07:14:47 +00:00
* feat: Add PDF import with AI-powered document analysis This enhances the import functionality to support PDF files with AI-powered document analysis. When a PDF is uploaded, it is processed by AI to: - Identify the document type (bank statement, credit card statement, etc.) - Generate a summary of the document contents - Extract key metadata (institution, dates, balances, transaction count) After processing, an email is sent to the user asking for next steps. Key changes: - Add PdfImport model for handling PDF document imports - Add Provider::Openai::PdfProcessor for AI document analysis - Add ProcessPdfJob for async PDF processing - Add PdfImportMailer for user notification emails - Update imports controller to detect and handle PDF uploads - Add PDF import option to the new import page - Add i18n translations for all new strings - Add comprehensive tests for the new functionality * Add bank statement import with AI extraction - Create ImportBankStatement assistant function for MCP - Add BankStatementExtractor with chunked processing for small context windows - Register function in assistant configurable - Make PdfImport#pdf_file_content public for extractor access - Increase OpenAI request timeout to 600s for slow local models - Increase DB connection pool to 20 for concurrent operations Tested with M-Pesa bank statement via remote Ollama (qwen3:8b): - Successfully extracted 18 transactions - Generated CSV and created TransactionImport - Works with 3000 char chunks for small context windows * Add pdf-reader gem dependency The BankStatementExtractor uses PDF::Reader to parse bank statement PDFs, but the gem was not properly declared in the Gemfile. This would cause NameError in production when processing bank statements. Added pdf-reader ~> 2.12 to Gemfile dependencies. * Fix transaction deduplication to preserve legitimate duplicates The previous deduplication logic removed ALL duplicate transactions based on [date, amount, name], which would drop legitimate same-day duplicates like multiple ATM withdrawals or card authorizations. Changed to only deduplicate transactions that appear in consecutive chunks (chunking artifacts) while preserving all legitimate duplicates within the same chunk or non-adjacent chunks. * Refactor bank statement extraction to use public provider method Address code review feedback: - Add public extract_bank_statement method to Provider::Openai - Remove direct access to private client via send(:client) - Update ImportBankStatement to use new public method - Add require 'set' to BankStatementExtractor - Remove PII-sensitive content from error logs - Add defensive check for nil response.error - Handle oversized PDF pages in chunking logic - Remove unused process_native and process_generic methods - Update email copy to reflect feature availability - Add guard for nil document_type in email template - Document pdf-reader gem rationale in Gemfile Tested with both OpenAI (gpt-4o) and Ollama (qwen3:8b): - OpenAI: 49 transactions extracted in 30s - Ollama: 40 transactions extracted in 368s - All encapsulation and error handling working correctly * Update schema.rb with ai_summary and document_type columns * Address PR #808 review comments - Rename :csv_file to :import_file across controllers/views/tests - Add PDF test fixture (sample_bank_statement.pdf) - Add supports_pdf_processing? method for graceful degradation - Revert unrelated database.yml pool change (600->3) - Remove month_start_day schema bleed from other PR - Fix PdfProcessor: use .strip instead of .strip_heredoc - Add server-side PDF magic byte validation - Conditionally show PDF import option when AI provider available - Fix ProcessPdfJob: sanitize errors, handle update failure - Move pdf_file attachment from Import to PdfImport - Document deduplication logic limitations - Fix ImportBankStatement: catch specific exceptions only - Remove unnecessary require 'set' - Remove dead json_schema method from PdfProcessor - Reduce default OpenAI timeout from 600s to 60s - Fix nil guard in text mailer template - Add require 'csv' to ImportBankStatement - Remove Gemfile pdf-reader comment * Fix RuboCop indentation in ProcessPdfJob * Refactor PDF import check to use model predicate method Replace is_a?(PdfImport) type check with requires_csv_workflow? predicate that leverages STI inheritance for cleaner controller logic. * Fix missing 'unknown' locale key and schema version mismatch - Add 'unknown: Unknown Document' to document_types locale - Fix schema version to match latest migration (2026_01_24_180211) * Document OPENAI_REQUEST_TIMEOUT env variable Added to .env.local.example and docs/hosting/ai.md * Rename ALLOWED_MIME_TYPES to ALLOWED_CSV_MIME_TYPES for clarity * Add comment explaining requires_csv_workflow? predicate * Remove redundant required_column_keys from PdfImport Base class already returns [] by default * Add ENV toggle to disable PDF processing for non-vision endpoints OPENAI_SUPPORTS_PDF_PROCESSING=false can be used for OpenAI-compatible endpoints (e.g., Ollama) that don't support vision/PDF processing. * Wire up transaction extraction for PDF bank statements - Add extracted_data JSONB column to imports - Add extract_transactions method to PdfImport - Call extraction in ProcessPdfJob for bank statements - Store transactions in extracted_data for later review * Fix ProcessPdfJob retry logic, sanitize and localize errors - Allow retries after partial success (classification ok, extraction failed) - Log sanitized error message instead of raw message to avoid data leakage - Use i18n for user-facing error messages * Add vision-capable model validation for PDF processing * Fix drag-and-drop test to use correct field name csv_file * Schema bleedover from another branch * Fix drag-drop import form field name to match controller * Add vision capability guard to process_pdf method --------- Co-authored-by: Claude <noreply@anthropic.com> Co-authored-by: mkdev11 <jaysmth689+github@users.noreply.github.com> Co-authored-by: Juan José Mata <jjmata@jjmata.com>
214 lines
7.1 KiB
Ruby
214 lines
7.1 KiB
Ruby
class Provider::Openai::BankStatementExtractor
|
|
MAX_CHARS_PER_CHUNK = 3000
|
|
attr_reader :client, :pdf_content, :model
|
|
|
|
def initialize(client:, pdf_content:, model:)
|
|
@client = client
|
|
@pdf_content = pdf_content
|
|
@model = model
|
|
end
|
|
|
|
def extract
|
|
pages = extract_pages_from_pdf
|
|
raise Provider::Openai::Error, "Could not extract text from PDF" if pages.empty?
|
|
|
|
chunks = build_chunks(pages)
|
|
Rails.logger.info("BankStatementExtractor: Processing #{chunks.size} chunk(s) from #{pages.size} page(s)")
|
|
|
|
all_transactions = []
|
|
metadata = {}
|
|
|
|
chunks.each_with_index do |chunk, index|
|
|
Rails.logger.info("BankStatementExtractor: Processing chunk #{index + 1}/#{chunks.size}")
|
|
result = process_chunk(chunk, index == 0)
|
|
|
|
# Tag transactions with chunk index for deduplication
|
|
tagged_transactions = (result[:transactions] || []).map { |t| t.merge(chunk_index: index) }
|
|
all_transactions.concat(tagged_transactions)
|
|
|
|
if index == 0
|
|
metadata = {
|
|
account_holder: result[:account_holder],
|
|
account_number: result[:account_number],
|
|
bank_name: result[:bank_name],
|
|
opening_balance: result[:opening_balance],
|
|
closing_balance: result[:closing_balance],
|
|
period: result[:period]
|
|
}
|
|
end
|
|
|
|
if result[:closing_balance].present?
|
|
metadata[:closing_balance] = result[:closing_balance]
|
|
end
|
|
if result.dig(:period, :end_date).present?
|
|
metadata[:period] ||= {}
|
|
metadata[:period][:end_date] = result.dig(:period, :end_date)
|
|
end
|
|
end
|
|
|
|
{
|
|
transactions: deduplicate_transactions(all_transactions),
|
|
period: metadata[:period] || {},
|
|
account_holder: metadata[:account_holder],
|
|
account_number: metadata[:account_number],
|
|
bank_name: metadata[:bank_name],
|
|
opening_balance: metadata[:opening_balance],
|
|
closing_balance: metadata[:closing_balance]
|
|
}
|
|
end
|
|
|
|
private
|
|
|
|
def extract_pages_from_pdf
|
|
return [] if pdf_content.blank?
|
|
|
|
reader = PDF::Reader.new(StringIO.new(pdf_content))
|
|
reader.pages.map(&:text).reject(&:blank?)
|
|
rescue => e
|
|
Rails.logger.error("Failed to extract text from PDF: #{e.message}")
|
|
[]
|
|
end
|
|
|
|
def build_chunks(pages)
|
|
chunks = []
|
|
current_chunk = []
|
|
current_size = 0
|
|
|
|
pages.each do |page_text|
|
|
if page_text.length > MAX_CHARS_PER_CHUNK
|
|
chunks << current_chunk.join("\n\n") if current_chunk.any?
|
|
current_chunk = []
|
|
current_size = 0
|
|
chunks << page_text
|
|
next
|
|
end
|
|
|
|
if current_size + page_text.length > MAX_CHARS_PER_CHUNK && current_chunk.any?
|
|
chunks << current_chunk.join("\n\n")
|
|
current_chunk = []
|
|
current_size = 0
|
|
end
|
|
|
|
current_chunk << page_text
|
|
current_size += page_text.length
|
|
end
|
|
|
|
chunks << current_chunk.join("\n\n") if current_chunk.any?
|
|
chunks
|
|
end
|
|
|
|
def process_chunk(text, is_first_chunk)
|
|
params = {
|
|
model: model,
|
|
messages: [
|
|
{ role: "system", content: is_first_chunk ? instructions_with_metadata : instructions_transactions_only },
|
|
{ role: "user", content: "Extract transactions:\n\n#{text}" }
|
|
],
|
|
response_format: { type: "json_object" }
|
|
}
|
|
|
|
response = client.chat(parameters: params)
|
|
content = response.dig("choices", 0, "message", "content")
|
|
|
|
raise Provider::Openai::Error, "No response from AI" if content.blank?
|
|
|
|
parsed = parse_json_response(content)
|
|
|
|
{
|
|
transactions: normalize_transactions(parsed["transactions"] || []),
|
|
period: {
|
|
start_date: parsed.dig("statement_period", "start_date"),
|
|
end_date: parsed.dig("statement_period", "end_date")
|
|
},
|
|
account_holder: parsed["account_holder"],
|
|
account_number: parsed["account_number"],
|
|
bank_name: parsed["bank_name"],
|
|
opening_balance: parsed["opening_balance"],
|
|
closing_balance: parsed["closing_balance"]
|
|
}
|
|
end
|
|
|
|
def parse_json_response(content)
|
|
cleaned = content.gsub(%r{^```json\s*}i, "").gsub(/```\s*$/, "").strip
|
|
JSON.parse(cleaned)
|
|
rescue JSON::ParserError => e
|
|
Rails.logger.error("BankStatementExtractor JSON parse error: #{e.message} (content_length=#{content.to_s.bytesize})")
|
|
{ "transactions" => [] }
|
|
end
|
|
|
|
def deduplicate_transactions(transactions)
|
|
# Deduplicates transactions that appear in consecutive chunks (chunking artifacts).
|
|
#
|
|
# KNOWN LIMITATION: Legitimate duplicate transactions (same date, amount, merchant)
|
|
# that happen to appear in adjacent chunks will be incorrectly deduplicated.
|
|
# This is an acceptable trade-off since chunking artifacts are more common than
|
|
# true same-day duplicates at chunk boundaries. Transactions within the same
|
|
# chunk are always preserved regardless of similarity.
|
|
seen = Set.new
|
|
transactions.select do |t|
|
|
# Create key without chunk_index for deduplication
|
|
key = [ t[:date], t[:amount], t[:name], t[:chunk_index] ]
|
|
|
|
# Check if we've seen this exact transaction in a different chunk
|
|
duplicate = seen.any? do |prev_key|
|
|
prev_key[0..2] == key[0..2] && (prev_key[3] - key[3]).abs <= 1
|
|
end
|
|
|
|
seen << key
|
|
!duplicate
|
|
end.map { |t| t.except(:chunk_index) }
|
|
end
|
|
|
|
def normalize_transactions(transactions)
|
|
transactions.map do |txn|
|
|
{
|
|
date: parse_date(txn["date"]),
|
|
amount: parse_amount(txn["amount"]),
|
|
name: txn["description"] || txn["name"] || txn["merchant"],
|
|
category: infer_category(txn),
|
|
notes: txn["reference"] || txn["notes"]
|
|
}
|
|
end.compact
|
|
end
|
|
|
|
def parse_date(date_str)
|
|
return nil if date_str.blank?
|
|
|
|
Date.parse(date_str).strftime("%Y-%m-%d")
|
|
rescue ArgumentError
|
|
nil
|
|
end
|
|
|
|
def parse_amount(amount)
|
|
return nil if amount.nil?
|
|
|
|
if amount.is_a?(Numeric)
|
|
amount.to_f
|
|
else
|
|
amount.to_s.gsub(/[^0-9.\-]/, "").to_f
|
|
end
|
|
end
|
|
|
|
def infer_category(txn)
|
|
txn["category"] || txn["type"]
|
|
end
|
|
|
|
def instructions_with_metadata
|
|
<<~INSTRUCTIONS.strip
|
|
Extract bank statement data as JSON. Return:
|
|
{"bank_name":"...","account_holder":"...","account_number":"last 4 digits","statement_period":{"start_date":"YYYY-MM-DD","end_date":"YYYY-MM-DD"},"opening_balance":0.00,"closing_balance":0.00,"transactions":[{"date":"YYYY-MM-DD","description":"...","amount":-0.00}]}
|
|
|
|
Rules: Negative amounts for debits/expenses, positive for credits/deposits. Dates as YYYY-MM-DD. Extract ALL transactions. JSON only, no markdown.
|
|
INSTRUCTIONS
|
|
end
|
|
|
|
def instructions_transactions_only
|
|
<<~INSTRUCTIONS.strip
|
|
Extract transactions from bank statement text as JSON. Return:
|
|
{"transactions":[{"date":"YYYY-MM-DD","description":"...","amount":-0.00}]}
|
|
|
|
Rules: Negative amounts for debits/expenses, positive for credits/deposits. Dates as YYYY-MM-DD. Extract ALL transactions. JSON only, no markdown.
|
|
INSTRUCTIONS
|
|
end
|
|
end
|