Compare commits
9 Commits
d60da0a7aa
...
master
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
adfcd90dcf | ||
|
|
bb610e0480 | ||
|
|
44d6cf6352 | ||
|
|
74529d8b19 | ||
|
|
bd7c6dddbf | ||
|
|
8e8d93e4a6 | ||
|
|
fee9f02faa | ||
|
|
bcd7eea1b4 | ||
|
|
281754de22 |
@@ -1,6 +1,4 @@
|
||||
---
|
||||
alwaysApply: true
|
||||
---
|
||||
sqlcmd -C -S tcp:192.168.56.1,1497 -U app -P 'readonly' -d eazybusiness -W
|
||||
|
||||
sqlcmd -C -S tcp:192.168.56.1,1497 -U sa -P 'sa_tekno23' -d eazybusiness -W
|
||||
7
.cursor/rules/devserver.mdc
Normal file
7
.cursor/rules/devserver.mdc
Normal file
@@ -0,0 +1,7 @@
|
||||
---
|
||||
alwaysApply: true
|
||||
---
|
||||
pm2 restart 10 -> restart backend (configured as "npm run dev:backend")
|
||||
pm2 restart 11 -> restart backend (configured as "npm run dev:frontend")
|
||||
|
||||
(both should rarely neer restart because in dev mode HMR for frontend, and nodemon for backend should already do that)
|
||||
@@ -8,6 +8,9 @@ REACT_APP_GOOGLE_CLIENT_ID=your_google_client_id_here
|
||||
# JWT Secret
|
||||
JWT_SECRET=your_jwt_secret_here
|
||||
|
||||
# OpenAI API Configuration
|
||||
OPENAI_API_KEY=your_openai_api_key_here
|
||||
|
||||
# Authorized Email Addresses (comma-separated)
|
||||
AUTHORIZED_EMAILS=admin@example.com,user1@example.com,user2@example.com
|
||||
|
||||
|
||||
@@ -1,14 +1,20 @@
|
||||
import React, { Component } from 'react';
|
||||
import { ThemeProvider, createTheme } from '@mui/material/styles';
|
||||
import CssBaseline from '@mui/material/CssBaseline';
|
||||
import { Container, AppBar, Toolbar, Typography, Button, Box, Tabs, Tab } from '@mui/material';
|
||||
import { Container, AppBar, Toolbar, Typography, Button, Box, Tabs, Tab, Badge, Chip, Divider, Snackbar, Alert, LinearProgress, Tooltip, CircularProgress } from '@mui/material';
|
||||
import LoginIcon from '@mui/icons-material/Login';
|
||||
import DashboardIcon from '@mui/icons-material/Dashboard';
|
||||
import DownloadIcon from '@mui/icons-material/Download';
|
||||
import TableChart from '@mui/icons-material/TableChart';
|
||||
import PlayArrowIcon from '@mui/icons-material/PlayArrow';
|
||||
import DocumentScannerIcon from '@mui/icons-material/DocumentScanner';
|
||||
import ExtractIcon from '@mui/icons-material/TextSnippet';
|
||||
import EmailIcon from '@mui/icons-material/Email';
|
||||
import UploadIcon from '@mui/icons-material/Upload';
|
||||
import AuthService from './services/AuthService';
|
||||
import DataViewer from './components/DataViewer';
|
||||
import Login from './components/Login';
|
||||
import OAuthCallback from './components/OAuthCallback';
|
||||
|
||||
const theme = createTheme({
|
||||
palette: {
|
||||
@@ -31,6 +37,18 @@ class App extends Component {
|
||||
loading: true,
|
||||
exportData: null, // { selectedMonth, canExport, onExport }
|
||||
currentView: 'dashboard', // 'dashboard' or 'tables'
|
||||
documentStatus: null,
|
||||
processingStatus: {
|
||||
markdown: false,
|
||||
extraction: false,
|
||||
datevSync: false,
|
||||
datevUpload: false
|
||||
},
|
||||
snackbar: {
|
||||
open: false,
|
||||
message: '',
|
||||
severity: 'info' // 'success', 'error', 'warning', 'info'
|
||||
}
|
||||
};
|
||||
this.authService = new AuthService();
|
||||
}
|
||||
@@ -39,6 +57,15 @@ class App extends Component {
|
||||
this.checkAuthStatus();
|
||||
}
|
||||
|
||||
componentDidUpdate(prevState) {
|
||||
// Clear targetTab after navigation is complete
|
||||
if (this.state.targetTab && prevState.currentView !== this.state.currentView) {
|
||||
setTimeout(() => {
|
||||
this.setState({ targetTab: null });
|
||||
}, 100); // Small delay to ensure navigation completes
|
||||
}
|
||||
}
|
||||
|
||||
checkAuthStatus = async () => {
|
||||
try {
|
||||
const token = localStorage.getItem('token');
|
||||
@@ -46,6 +73,7 @@ class App extends Component {
|
||||
const user = await this.authService.verifyToken(token);
|
||||
if (user) {
|
||||
this.setState({ isAuthenticated: true, user, loading: false });
|
||||
this.fetchDocumentStatus();
|
||||
return;
|
||||
}
|
||||
}
|
||||
@@ -62,6 +90,7 @@ class App extends Component {
|
||||
if (result.success) {
|
||||
localStorage.setItem('token', result.token);
|
||||
this.setState({ isAuthenticated: true, user: result.user });
|
||||
this.fetchDocumentStatus();
|
||||
}
|
||||
} catch (error) {
|
||||
console.error('Login failed:', error);
|
||||
@@ -83,8 +112,135 @@ class App extends Component {
|
||||
this.setState({ currentView: newValue });
|
||||
};
|
||||
|
||||
showSnackbar = (message, severity = 'info') => {
|
||||
this.setState({
|
||||
snackbar: {
|
||||
open: true,
|
||||
message,
|
||||
severity
|
||||
}
|
||||
});
|
||||
};
|
||||
|
||||
handleSnackbarClose = (event, reason) => {
|
||||
if (reason === 'clickaway') {
|
||||
return;
|
||||
}
|
||||
this.setState({
|
||||
snackbar: {
|
||||
...this.state.snackbar,
|
||||
open: false
|
||||
}
|
||||
});
|
||||
};
|
||||
|
||||
fetchDocumentStatus = async () => {
|
||||
try {
|
||||
const token = localStorage.getItem('token');
|
||||
if (!token) {
|
||||
console.log('No token found for document status');
|
||||
return;
|
||||
}
|
||||
|
||||
console.log('Fetching document status...');
|
||||
const response = await fetch('/api/data/document-status', {
|
||||
headers: {
|
||||
'Authorization': `Bearer ${token}`,
|
||||
'Content-Type': 'application/json',
|
||||
},
|
||||
});
|
||||
|
||||
if (response.ok) {
|
||||
const status = await response.json();
|
||||
console.log('Document status received:', status);
|
||||
this.setState({ documentStatus: status });
|
||||
} else {
|
||||
console.error('Failed to fetch document status:', response.status, await response.text());
|
||||
}
|
||||
} catch (error) {
|
||||
console.error('Error fetching document status:', error);
|
||||
}
|
||||
};
|
||||
|
||||
handleProcessing = async (processType) => {
|
||||
if (this.state.processingStatus[processType]) {
|
||||
return; // Already processing
|
||||
}
|
||||
|
||||
// Handle datev upload navigation
|
||||
if (processType === 'datev-upload') {
|
||||
this.setState({
|
||||
currentView: 'tables',
|
||||
targetTab: {
|
||||
level1: 3, // CSV Import tab
|
||||
level2: 'DATEV_LINKS' // DATEV Beleglinks tab
|
||||
}
|
||||
});
|
||||
return;
|
||||
}
|
||||
|
||||
// Check if there are documents to process
|
||||
const statusKey = processType === 'datev-sync' ? 'needDatevSync' :
|
||||
processType === 'extraction' ? 'needExtraction' : 'needMarkdown';
|
||||
|
||||
if (!this.state.documentStatus || this.state.documentStatus[statusKey] === 0) {
|
||||
this.showSnackbar(`No documents need ${processType} processing at this time.`, 'info');
|
||||
return;
|
||||
}
|
||||
|
||||
this.setState(prevState => ({
|
||||
processingStatus: {
|
||||
...prevState.processingStatus,
|
||||
[processType]: true
|
||||
}
|
||||
}));
|
||||
|
||||
try {
|
||||
const token = localStorage.getItem('token');
|
||||
if (!token) return;
|
||||
|
||||
const response = await fetch(`/api/data/process-${processType}`, {
|
||||
method: 'POST',
|
||||
headers: {
|
||||
'Authorization': `Bearer ${token}`,
|
||||
'Content-Type': 'application/json',
|
||||
},
|
||||
});
|
||||
|
||||
if (response.ok) {
|
||||
const result = await response.json();
|
||||
console.log(`${processType} processing result:`, result);
|
||||
this.showSnackbar(`${processType} processing completed successfully!`, 'success');
|
||||
// Refresh document status after successful processing
|
||||
await this.fetchDocumentStatus();
|
||||
} else {
|
||||
const error = await response.json();
|
||||
console.error(`Failed to process ${processType}:`, error);
|
||||
this.showSnackbar(`Failed to process ${processType}: ${error.error || response.status}`, 'error');
|
||||
}
|
||||
} catch (error) {
|
||||
console.error(`Error processing ${processType}:`, error);
|
||||
this.showSnackbar(`Error processing ${processType}: ${error.message}`, 'error');
|
||||
} finally {
|
||||
this.setState(prevState => ({
|
||||
processingStatus: {
|
||||
...prevState.processingStatus,
|
||||
[processType]: false
|
||||
}
|
||||
}));
|
||||
}
|
||||
};
|
||||
|
||||
isOAuthCallback = () => {
|
||||
return window.location.pathname === '/auth/callback';
|
||||
};
|
||||
|
||||
render() {
|
||||
const { isAuthenticated, user, loading, currentView } = this.state;
|
||||
const { isAuthenticated, user, loading, currentView, documentStatus, processingStatus, snackbar } = this.state;
|
||||
|
||||
// Debug logging
|
||||
console.log('App render - documentStatus:', documentStatus);
|
||||
console.log('App render - isAuthenticated:', isAuthenticated);
|
||||
|
||||
if (loading) {
|
||||
return (
|
||||
@@ -149,6 +305,112 @@ class App extends Component {
|
||||
sx={{ minHeight: 48 }}
|
||||
/>
|
||||
</Tabs>
|
||||
<Divider orientation="vertical" flexItem sx={{ mx: 2, backgroundColor: 'rgba(255, 255, 255, 0.3)' }} />
|
||||
<Box sx={{ display: 'flex', alignItems: 'center', gap: 1 }}>
|
||||
<Tooltip title={processingStatus.markdown ? 'Running markdown conversion… this can take a while' : 'Process markdown conversion'} arrow>
|
||||
<span>
|
||||
<Button
|
||||
color="inherit"
|
||||
size="small"
|
||||
onClick={() => this.handleProcessing('markdown')}
|
||||
disabled={processingStatus.markdown || !documentStatus}
|
||||
sx={{
|
||||
minWidth: 'auto',
|
||||
px: 1,
|
||||
'&:hover': { backgroundColor: 'rgba(255, 255, 255, 0.1)' }
|
||||
}}
|
||||
>
|
||||
<Badge
|
||||
badgeContent={documentStatus?.needMarkdown || 0}
|
||||
color={documentStatus?.needMarkdown > 0 ? "error" : "default"}
|
||||
max={999999}
|
||||
sx={{ mr: 0.5 }}
|
||||
>
|
||||
<DocumentScannerIcon fontSize="small" />
|
||||
</Badge>
|
||||
{processingStatus.markdown && (
|
||||
<CircularProgress size={14} color="inherit" />
|
||||
)}
|
||||
</Button>
|
||||
</span>
|
||||
</Tooltip>
|
||||
<Tooltip title={processingStatus.extraction ? 'Running data extraction… this can take a while' : 'Process data extraction'} arrow>
|
||||
<span>
|
||||
<Button
|
||||
color="inherit"
|
||||
size="small"
|
||||
onClick={() => this.handleProcessing('extraction')}
|
||||
disabled={processingStatus.extraction || !documentStatus}
|
||||
sx={{
|
||||
minWidth: 'auto',
|
||||
px: 1,
|
||||
'&:hover': { backgroundColor: 'rgba(255, 255, 255, 0.1)' }
|
||||
}}
|
||||
>
|
||||
<Badge
|
||||
badgeContent={documentStatus?.needExtraction || 0}
|
||||
color={documentStatus?.needExtraction > 0 ? "warning" : "default"}
|
||||
max={999999}
|
||||
sx={{ mr: 0.5 }}
|
||||
>
|
||||
<ExtractIcon fontSize="small" />
|
||||
</Badge>
|
||||
{processingStatus.extraction && (
|
||||
<CircularProgress size={14} color="inherit" />
|
||||
)}
|
||||
</Button>
|
||||
</span>
|
||||
</Tooltip>
|
||||
<Tooltip title={processingStatus.datevSync ? 'Running DATEV sync… this can take a while' : 'Process Datev sync'} arrow>
|
||||
<span>
|
||||
<Button
|
||||
color="inherit"
|
||||
size="small"
|
||||
onClick={() => this.handleProcessing('datev-sync')}
|
||||
disabled={processingStatus.datevSync || !documentStatus}
|
||||
sx={{
|
||||
minWidth: 'auto',
|
||||
px: 1,
|
||||
'&:hover': { backgroundColor: 'rgba(255, 255, 255, 0.1)' }
|
||||
}}
|
||||
>
|
||||
<Badge
|
||||
badgeContent={documentStatus?.needDatevSync || 0}
|
||||
color={documentStatus?.needDatevSync > 0 ? "info" : "default"}
|
||||
max={999999}
|
||||
sx={{ mr: 0.5 }}
|
||||
>
|
||||
<EmailIcon fontSize="small" />
|
||||
</Badge>
|
||||
{processingStatus.datevSync && (
|
||||
<CircularProgress size={14} color="inherit" />
|
||||
)}
|
||||
</Button>
|
||||
</span>
|
||||
</Tooltip>
|
||||
<Button
|
||||
color="inherit"
|
||||
size="small"
|
||||
onClick={() => this.handleProcessing('datev-upload')}
|
||||
disabled={processingStatus.datevUpload || !documentStatus}
|
||||
sx={{
|
||||
minWidth: 'auto',
|
||||
px: 1,
|
||||
'&:hover': { backgroundColor: 'rgba(255, 255, 255, 0.1)' }
|
||||
}}
|
||||
title="Process Datev CSV upload"
|
||||
>
|
||||
<Badge
|
||||
badgeContent={documentStatus?.needDatevUpload || 0}
|
||||
color={documentStatus?.needDatevUpload > 0 ? "secondary" : "default"}
|
||||
max={999999}
|
||||
sx={{ mr: 0.5 }}
|
||||
>
|
||||
<UploadIcon fontSize="small" />
|
||||
</Badge>
|
||||
{processingStatus.datevUpload && <PlayArrowIcon fontSize="small" />}
|
||||
</Button>
|
||||
</Box>
|
||||
{this.state.exportData && (
|
||||
<Button
|
||||
color="inherit"
|
||||
@@ -184,22 +446,44 @@ class App extends Component {
|
||||
</>
|
||||
)}
|
||||
</Toolbar>
|
||||
{(processingStatus.markdown || processingStatus.extraction || processingStatus.datevSync) && (
|
||||
<LinearProgress color="secondary" />
|
||||
)}
|
||||
</AppBar>
|
||||
|
||||
<Box sx={{ height: 'calc(100vh - 64px)', display: 'flex', flexDirection: 'column' }}>
|
||||
<Container maxWidth={false} sx={{ mt: 4, flex: 1, minHeight: 0, display: 'flex', flexDirection: 'column', width: '100%' }}>
|
||||
{isAuthenticated ? (
|
||||
{this.isOAuthCallback() ? (
|
||||
<OAuthCallback />
|
||||
) : isAuthenticated ? (
|
||||
<DataViewer
|
||||
user={user}
|
||||
onUpdateExportData={this.updateExportData}
|
||||
currentView={currentView}
|
||||
onViewChange={this.handleViewChange}
|
||||
targetTab={this.state.targetTab}
|
||||
/>
|
||||
) : (
|
||||
<Login onLogin={this.handleLogin} />
|
||||
)}
|
||||
</Container>
|
||||
</Box>
|
||||
|
||||
<Snackbar
|
||||
open={snackbar.open}
|
||||
autoHideDuration={6000}
|
||||
onClose={this.handleSnackbarClose}
|
||||
anchorOrigin={{ vertical: 'bottom', horizontal: 'right' }}
|
||||
>
|
||||
<Alert
|
||||
onClose={this.handleSnackbarClose}
|
||||
severity={snackbar.severity}
|
||||
variant="filled"
|
||||
sx={{ width: '100%' }}
|
||||
>
|
||||
{snackbar.message}
|
||||
</Alert>
|
||||
</Snackbar>
|
||||
</ThemeProvider>
|
||||
);
|
||||
}
|
||||
|
||||
@@ -44,6 +44,7 @@ class AccountingItemsManager extends Component {
|
||||
editingItem: null,
|
||||
showCreateDialog: false,
|
||||
showCreateKontoDialog: false,
|
||||
jtlKontierung: null,
|
||||
newItem: {
|
||||
umsatz_brutto: '',
|
||||
soll_haben_kz: 'S',
|
||||
@@ -65,6 +66,7 @@ class AccountingItemsManager extends Component {
|
||||
|
||||
componentDidMount() {
|
||||
this.loadData();
|
||||
this.loadJtlKontierung();
|
||||
}
|
||||
|
||||
loadData = async () => {
|
||||
@@ -88,6 +90,31 @@ class AccountingItemsManager extends Component {
|
||||
}
|
||||
};
|
||||
|
||||
loadJtlKontierung = async () => {
|
||||
try {
|
||||
const { transaction } = this.props;
|
||||
if (!transaction || !transaction.jtlId) {
|
||||
this.setState({ jtlKontierung: undefined });
|
||||
return;
|
||||
}
|
||||
|
||||
const response = await this.authService.apiCall(`/data/jtl-kontierung/${transaction.jtlId}`);
|
||||
if (!response) return;
|
||||
|
||||
if (response.ok) {
|
||||
const data = await response.json();
|
||||
this.setState({ jtlKontierung: data });
|
||||
} else {
|
||||
const err = await response.json();
|
||||
console.error('Failed to load JTL Kontierung:', err);
|
||||
this.setState({ jtlKontierung: undefined });
|
||||
}
|
||||
} catch (e) {
|
||||
console.error('Error loading JTL Kontierung:', e);
|
||||
this.setState({ jtlKontierung: undefined });
|
||||
}
|
||||
}
|
||||
|
||||
loadAccountingItems = async () => {
|
||||
const { transaction } = this.props;
|
||||
if (!transaction?.id) return;
|
||||
@@ -319,6 +346,21 @@ class AccountingItemsManager extends Component {
|
||||
/>
|
||||
</Box>
|
||||
|
||||
{transaction?.jtlId && (
|
||||
<Box sx={{ mb: 2, p: 2, border: '1px dashed #999', borderRadius: 1 }}>
|
||||
<Typography variant="subtitle2">Debug: tUmsatzKontierung.data</Typography>
|
||||
<Typography variant="caption" component="div" sx={{ whiteSpace: 'pre-wrap', wordBreak: 'break-word' }}>
|
||||
{this.state.jtlKontierung === undefined
|
||||
? 'undefined'
|
||||
: this.state.jtlKontierung === null
|
||||
? 'null'
|
||||
: typeof this.state.jtlKontierung === 'object'
|
||||
? JSON.stringify(this.state.jtlKontierung, null, 2)
|
||||
: String(this.state.jtlKontierung)}
|
||||
</Typography>
|
||||
</Box>
|
||||
)}
|
||||
|
||||
<TableContainer component={Paper}>
|
||||
<Table size="small">
|
||||
<TableHead>
|
||||
|
||||
@@ -56,6 +56,20 @@ class CSVImportPanel extends Component {
|
||||
this.datevFileInputRef = React.createRef();
|
||||
}
|
||||
|
||||
componentDidMount() {
|
||||
// Check if we should navigate to a specific tab
|
||||
if (this.props.targetTab) {
|
||||
this.setState({ activeTab: this.props.targetTab });
|
||||
}
|
||||
}
|
||||
|
||||
componentDidUpdate(prevProps) {
|
||||
// Handle targetTab changes
|
||||
if (this.props.targetTab !== prevProps.targetTab && this.props.targetTab) {
|
||||
this.setState({ activeTab: this.props.targetTab });
|
||||
}
|
||||
}
|
||||
|
||||
// Tab switch resets type-specific state but keeps success state as-is
|
||||
handleTabChange = (_e, value) => {
|
||||
this.setState({
|
||||
@@ -344,6 +358,7 @@ class CSVImportPanel extends Component {
|
||||
)}
|
||||
|
||||
{currentHeaders && (
|
||||
|
||||
<Box sx={{ mb: 2 }}>
|
||||
<Typography variant="subtitle2" gutterBottom>
|
||||
Erkannte Spalten ({currentHeaders.length}):
|
||||
@@ -442,7 +457,7 @@ class CSVImportPanel extends Component {
|
||||
)}
|
||||
{importResult.errors > 0 && (
|
||||
<Typography variant="body1" color="warning.main">
|
||||
<strong>Fehler:</strong> {importResult.errors} Zeilen konnten nicht verarbeitet werden
|
||||
<strong>Fehler:</strong> {importResult.errors} Zeilen konnten nicht verarbeitet werden.
|
||||
</Typography>
|
||||
)}
|
||||
{importResult.message && (
|
||||
|
||||
@@ -165,7 +165,7 @@ class DataViewer extends Component {
|
||||
</>
|
||||
) : (
|
||||
<Box sx={{ flex: 1, minHeight: 0, overflow: 'auto', p: 2 }}>
|
||||
<TableManagement user={user} />
|
||||
<TableManagement user={user} targetTab={this.props.targetTab} />
|
||||
</Box>
|
||||
)}
|
||||
</Box>
|
||||
|
||||
@@ -9,6 +9,10 @@ class Login extends Component {
|
||||
error: null,
|
||||
loading: false,
|
||||
};
|
||||
|
||||
// Flags to track FedCM attempts and success
|
||||
this.fedcmAttempted = false;
|
||||
this.fedcmSucceeded = false;
|
||||
}
|
||||
|
||||
componentDidMount() {
|
||||
@@ -34,11 +38,17 @@ class Login extends Component {
|
||||
initializeGoogleSignIn = () => {
|
||||
if (window.google && window.google.accounts) {
|
||||
try {
|
||||
// Note: Removed debug logging to avoid deprecated method warnings
|
||||
|
||||
console.log('REACT_APP_GOOGLE_CLIENT_ID', process.env.REACT_APP_GOOGLE_CLIENT_ID);
|
||||
console.log('Current origin for Google auth:', window.location.origin);
|
||||
console.log('User agent:', navigator.userAgent);
|
||||
|
||||
window.google.accounts.id.initialize({
|
||||
client_id: process.env.REACT_APP_GOOGLE_CLIENT_ID || 'your_google_client_id_here',
|
||||
client_id: process.env.REACT_APP_GOOGLE_CLIENT_ID,
|
||||
callback: this.handleGoogleResponse,
|
||||
auto_select: false,
|
||||
cancel_on_tap_outside: true,
|
||||
cancel_on_tap_outside: false,
|
||||
});
|
||||
console.log('✅ Google Sign-In initialized');
|
||||
} catch (error) {
|
||||
@@ -48,6 +58,9 @@ class Login extends Component {
|
||||
};
|
||||
|
||||
handleGoogleResponse = (response) => {
|
||||
// Mark FedCM as successful if we get here
|
||||
this.fedcmSucceeded = true;
|
||||
|
||||
this.setState({ loading: true, error: null });
|
||||
this.props.onLogin(response)
|
||||
.catch((error) => {
|
||||
@@ -70,6 +83,11 @@ class Login extends Component {
|
||||
errorMessage = '🚫 Zugriff verweigert: Ihre E-Mail-Adresse ist nicht autorisiert. Versuchen Sie, sich mit einem anderen Google-Konto anzumelden.';
|
||||
} else if (error.message.includes('No authorized users configured')) {
|
||||
errorMessage = '🔒 Kein Zugriff: Derzeit sind keine Benutzer autorisiert. Wenden Sie sich an den Administrator.';
|
||||
} else if (error.message.includes('Not signed in with the identity provider') ||
|
||||
error.message.includes('NetworkError') ||
|
||||
error.message.includes('FedCM')) {
|
||||
// FedCM failed, offer redirect option
|
||||
errorMessage = '🔄 Schnelle Anmeldung nicht verfügbar. Versuchen Sie die Standard-Anmeldung.';
|
||||
} else {
|
||||
// Show the actual error message from the server
|
||||
errorMessage = `❌ Anmeldefehler: ${error.message}`;
|
||||
@@ -92,29 +110,105 @@ class Login extends Component {
|
||||
return;
|
||||
}
|
||||
|
||||
// Clear any previous error
|
||||
this.setState({ error: null, loading: false });
|
||||
// Clear any previous error and start loading
|
||||
this.setState({ error: null, loading: true });
|
||||
|
||||
// Try FedCM first (seamless for users already signed in to Google)
|
||||
console.log('🎯 Trying FedCM first for optimal UX...');
|
||||
this.tryFedCMFirst();
|
||||
};
|
||||
|
||||
tryFedCMFirst = () => {
|
||||
if (window.google && window.google.accounts && window.google.accounts.id) {
|
||||
try {
|
||||
window.google.accounts.id.prompt();
|
||||
} catch (error) {
|
||||
console.error('Google prompt error:', error);
|
||||
this.setState({
|
||||
error: 'Google-Anmeldung konnte nicht geladen werden. Die Seite wird aktualisiert, um es erneut zu versuchen.',
|
||||
loading: true
|
||||
console.log('✅ Trying FedCM for seamless sign-in...');
|
||||
|
||||
// Listen for the specific FedCM errors that indicate no Google session
|
||||
const originalConsoleError = console.error;
|
||||
let errorIntercepted = false;
|
||||
|
||||
console.error = (...args) => {
|
||||
const errorMessage = args.join(' ');
|
||||
if (!errorIntercepted && (
|
||||
errorMessage.includes('Not signed in with the identity provider') ||
|
||||
errorMessage.includes('FedCM get() rejects with NetworkError') ||
|
||||
errorMessage.includes('Error retrieving a token')
|
||||
)) {
|
||||
errorIntercepted = true;
|
||||
console.error = originalConsoleError; // Restore immediately
|
||||
console.log('🔄 FedCM failed (user not signed in to Google), using redirect...');
|
||||
this.redirectToGoogleOAuth();
|
||||
return;
|
||||
}
|
||||
originalConsoleError.apply(console, args);
|
||||
};
|
||||
|
||||
// Try FedCM
|
||||
window.google.accounts.id.prompt((notification) => {
|
||||
console.log('🔍 FedCM notification:', notification);
|
||||
console.error = originalConsoleError; // Restore console.error
|
||||
// If we get here without error, FedCM is working
|
||||
});
|
||||
setTimeout(() => window.location.reload(), 2000);
|
||||
|
||||
} catch (error) {
|
||||
console.error('FedCM initialization error, falling back to redirect:', error);
|
||||
this.redirectToGoogleOAuth();
|
||||
}
|
||||
} else {
|
||||
this.setState({
|
||||
error: 'Google-Anmeldung nicht geladen. Die Seite wird aktualisiert, um es erneut zu versuchen.',
|
||||
loading: true
|
||||
});
|
||||
setTimeout(() => window.location.reload(), 2000);
|
||||
// Google Identity Services not loaded, go straight to redirect
|
||||
console.log('📋 GSI not loaded, using redirect flow...');
|
||||
this.redirectToGoogleOAuth();
|
||||
}
|
||||
};
|
||||
|
||||
redirectToGoogleOAuth = () => {
|
||||
try {
|
||||
// Generate a random state parameter for security
|
||||
const state = this.generateRandomString(32);
|
||||
sessionStorage.setItem('oauth_state', state);
|
||||
|
||||
// Build the Google OAuth2 authorization URL
|
||||
const params = new URLSearchParams({
|
||||
client_id: process.env.REACT_APP_GOOGLE_CLIENT_ID,
|
||||
redirect_uri: window.location.origin + '/auth/callback',
|
||||
response_type: 'code',
|
||||
scope: 'openid email profile',
|
||||
state: state,
|
||||
access_type: 'online',
|
||||
prompt: 'select_account'
|
||||
});
|
||||
|
||||
const authUrl = `https://accounts.google.com/o/oauth2/v2/auth?${params.toString()}`;
|
||||
|
||||
console.log('🔗 Redirecting to Google OAuth:', authUrl);
|
||||
|
||||
// Redirect to Google OAuth
|
||||
window.location.href = authUrl;
|
||||
|
||||
} catch (error) {
|
||||
console.error('Redirect OAuth error:', error);
|
||||
this.setState({
|
||||
error: 'Google-Anmeldung konnte nicht gestartet werden.',
|
||||
loading: false
|
||||
});
|
||||
}
|
||||
};
|
||||
|
||||
generateRandomString = (length) => {
|
||||
const chars = 'ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789';
|
||||
let result = '';
|
||||
for (let i = 0; i < length; i++) {
|
||||
result += chars.charAt(Math.floor(Math.random() * chars.length));
|
||||
}
|
||||
return result;
|
||||
};
|
||||
|
||||
handleUseRedirect = () => {
|
||||
console.log('🔄 User chose redirect flow');
|
||||
this.setState({ error: null, loading: true });
|
||||
this.redirectToGoogleOAuth();
|
||||
};
|
||||
|
||||
|
||||
|
||||
|
||||
@@ -157,6 +251,20 @@ class Login extends Component {
|
||||
{loading ? 'Anmeldung läuft...' : 'Mit Google anmelden'}
|
||||
</Button>
|
||||
|
||||
{error && error.includes('Standard-Anmeldung') && (
|
||||
<Button
|
||||
fullWidth
|
||||
variant="outlined"
|
||||
size="large"
|
||||
startIcon={<GoogleIcon />}
|
||||
onClick={this.handleUseRedirect}
|
||||
disabled={loading}
|
||||
sx={{ py: 1.5, mt: 2 }}
|
||||
>
|
||||
Standard Google-Anmeldung verwenden
|
||||
</Button>
|
||||
)}
|
||||
|
||||
|
||||
|
||||
<Typography variant="caption" display="block" textAlign="center" sx={{ mt: 2 }}>
|
||||
|
||||
133
client/src/components/OAuthCallback.js
Normal file
133
client/src/components/OAuthCallback.js
Normal file
@@ -0,0 +1,133 @@
|
||||
import React, { Component } from 'react';
|
||||
import { Box, CircularProgress, Typography, Alert } from '@mui/material';
|
||||
|
||||
class OAuthCallback extends Component {
|
||||
constructor(props) {
|
||||
super(props);
|
||||
this.state = {
|
||||
loading: true,
|
||||
error: null,
|
||||
};
|
||||
}
|
||||
|
||||
componentDidMount() {
|
||||
this.handleOAuthCallback();
|
||||
}
|
||||
|
||||
handleOAuthCallback = async () => {
|
||||
try {
|
||||
const urlParams = new URLSearchParams(window.location.search);
|
||||
const code = urlParams.get('code');
|
||||
const state = urlParams.get('state');
|
||||
const error = urlParams.get('error');
|
||||
|
||||
// Check for OAuth errors
|
||||
if (error) {
|
||||
throw new Error(`OAuth error: ${error}`);
|
||||
}
|
||||
|
||||
// Verify state parameter for security
|
||||
const storedState = sessionStorage.getItem('oauth_state');
|
||||
if (!state || state !== storedState) {
|
||||
throw new Error('Invalid state parameter - possible CSRF attack');
|
||||
}
|
||||
|
||||
// Clear stored state
|
||||
sessionStorage.removeItem('oauth_state');
|
||||
|
||||
if (!code) {
|
||||
throw new Error('No authorization code received');
|
||||
}
|
||||
|
||||
console.log('🔑 Authorization code received, exchanging for tokens...');
|
||||
|
||||
// Exchange authorization code for tokens via our backend
|
||||
const response = await fetch('/api/auth/google/callback', {
|
||||
method: 'POST',
|
||||
headers: {
|
||||
'Content-Type': 'application/json',
|
||||
},
|
||||
body: JSON.stringify({
|
||||
code: code,
|
||||
redirect_uri: window.location.origin + '/auth/callback'
|
||||
}),
|
||||
});
|
||||
|
||||
if (!response.ok) {
|
||||
const errorData = await response.json().catch(() => ({}));
|
||||
throw new Error(errorData.message || `HTTP ${response.status}: ${response.statusText}`);
|
||||
}
|
||||
|
||||
const data = await response.json();
|
||||
|
||||
if (data.success && data.token) {
|
||||
console.log('✅ OAuth callback successful');
|
||||
|
||||
// Store the JWT token
|
||||
localStorage.setItem('token', data.token);
|
||||
|
||||
// Redirect to main app
|
||||
window.location.href = '/';
|
||||
} else {
|
||||
throw new Error(data.message || 'Authentication failed');
|
||||
}
|
||||
|
||||
} catch (error) {
|
||||
console.error('OAuth callback error:', error);
|
||||
this.setState({
|
||||
loading: false,
|
||||
error: error.message || 'Authentication failed'
|
||||
});
|
||||
}
|
||||
};
|
||||
|
||||
render() {
|
||||
const { loading, error } = this.state;
|
||||
|
||||
if (error) {
|
||||
return (
|
||||
<Box
|
||||
display="flex"
|
||||
justifyContent="center"
|
||||
alignItems="center"
|
||||
minHeight="60vh"
|
||||
flexDirection="column"
|
||||
>
|
||||
<Alert severity="error" sx={{ mb: 2, maxWidth: 400 }}>
|
||||
<Typography variant="h6" gutterBottom>
|
||||
Anmeldung fehlgeschlagen
|
||||
</Typography>
|
||||
<Typography variant="body2">
|
||||
{error}
|
||||
</Typography>
|
||||
</Alert>
|
||||
<Typography variant="body2" color="textSecondary">
|
||||
<a href="/" style={{ color: 'inherit' }}>
|
||||
Zurück zur Anmeldung
|
||||
</a>
|
||||
</Typography>
|
||||
</Box>
|
||||
);
|
||||
}
|
||||
|
||||
return (
|
||||
<Box
|
||||
display="flex"
|
||||
justifyContent="center"
|
||||
alignItems="center"
|
||||
minHeight="60vh"
|
||||
flexDirection="column"
|
||||
>
|
||||
<CircularProgress size={60} sx={{ mb: 2 }} />
|
||||
<Typography variant="h6" gutterBottom>
|
||||
Anmeldung wird verarbeitet...
|
||||
</Typography>
|
||||
<Typography variant="body2" color="textSecondary">
|
||||
Sie werden automatisch weitergeleitet.
|
||||
</Typography>
|
||||
</Box>
|
||||
);
|
||||
}
|
||||
}
|
||||
|
||||
export default OAuthCallback;
|
||||
@@ -26,6 +26,21 @@ class TableManagement extends Component {
|
||||
};
|
||||
}
|
||||
|
||||
componentDidMount() {
|
||||
// Check if we should navigate to a specific tab
|
||||
if (this.props.targetTab?.level1 !== undefined) {
|
||||
this.setState({ activeTab: this.props.targetTab.level1 });
|
||||
}
|
||||
}
|
||||
|
||||
componentDidUpdate(prevProps) {
|
||||
// Handle targetTab changes
|
||||
if (this.props.targetTab?.level1 !== prevProps.targetTab?.level1 &&
|
||||
this.props.targetTab?.level1 !== undefined) {
|
||||
this.setState({ activeTab: this.props.targetTab.level1 });
|
||||
}
|
||||
}
|
||||
|
||||
handleTabChange = (event, newValue) => {
|
||||
this.setState({ activeTab: newValue });
|
||||
};
|
||||
@@ -92,6 +107,7 @@ class TableManagement extends Component {
|
||||
</Typography>
|
||||
<CSVImportPanel
|
||||
user={user}
|
||||
targetTab={this.props.targetTab?.level2}
|
||||
/>
|
||||
</Box>
|
||||
)}
|
||||
|
||||
@@ -1,56 +0,0 @@
|
||||
version: '3.8'
|
||||
|
||||
services:
|
||||
nginx:
|
||||
image: nginx:alpine
|
||||
ports:
|
||||
- "80:80"
|
||||
volumes:
|
||||
- ./nginx.dev.conf:/etc/nginx/conf.d/default.conf
|
||||
- ./logs/nginx:/var/log/nginx
|
||||
depends_on:
|
||||
- frontend
|
||||
- backend
|
||||
restart: unless-stopped
|
||||
networks:
|
||||
- fibdash-network
|
||||
|
||||
frontend:
|
||||
build:
|
||||
context: .
|
||||
dockerfile: Dockerfile.dev.frontend
|
||||
ports:
|
||||
- "5001:5001"
|
||||
volumes:
|
||||
- ./client:/app/client
|
||||
- /app/node_modules
|
||||
environment:
|
||||
- NODE_ENV=development
|
||||
- CHOKIDAR_USEPOLLING=true
|
||||
networks:
|
||||
- fibdash-network
|
||||
command: npm run dev:frontend
|
||||
|
||||
backend:
|
||||
build:
|
||||
context: .
|
||||
dockerfile: Dockerfile.dev.backend
|
||||
ports:
|
||||
- "5000:5000"
|
||||
volumes:
|
||||
- ./src:/app/src
|
||||
- /app/node_modules
|
||||
environment:
|
||||
- NODE_ENV=development
|
||||
env_file:
|
||||
- .env
|
||||
networks:
|
||||
- fibdash-network
|
||||
command: npm run dev:backend
|
||||
|
||||
networks:
|
||||
fibdash-network:
|
||||
driver: bridge
|
||||
|
||||
volumes:
|
||||
node_modules:
|
||||
34
package-lock.json
generated
34
package-lock.json
generated
@@ -21,6 +21,8 @@
|
||||
"google-auth-library": "^9.0.0",
|
||||
"jsonwebtoken": "^9.0.0",
|
||||
"mssql": "^9.1.0",
|
||||
"nodemailer": "^7.0.5",
|
||||
"openai": "^5.12.0",
|
||||
"react": "^18.2.0",
|
||||
"react-dom": "^18.2.0"
|
||||
},
|
||||
@@ -7142,6 +7144,15 @@
|
||||
"dev": true,
|
||||
"license": "MIT"
|
||||
},
|
||||
"node_modules/nodemailer": {
|
||||
"version": "7.0.5",
|
||||
"resolved": "https://registry.npmjs.org/nodemailer/-/nodemailer-7.0.5.tgz",
|
||||
"integrity": "sha512-nsrh2lO3j4GkLLXoeEksAMgAOqxOv6QumNRVQTJwKH4nuiww6iC2y7GyANs9kRAxCexg3+lTWM3PZ91iLlVjfg==",
|
||||
"license": "MIT-0",
|
||||
"engines": {
|
||||
"node": ">=6.0.0"
|
||||
}
|
||||
},
|
||||
"node_modules/nodemon": {
|
||||
"version": "3.1.10",
|
||||
"resolved": "https://registry.npmjs.org/nodemon/-/nodemon-3.1.10.tgz",
|
||||
@@ -7365,6 +7376,27 @@
|
||||
"url": "https://github.com/sponsors/sindresorhus"
|
||||
}
|
||||
},
|
||||
"node_modules/openai": {
|
||||
"version": "5.12.0",
|
||||
"resolved": "https://registry.npmjs.org/openai/-/openai-5.12.0.tgz",
|
||||
"integrity": "sha512-vUdt02xiWgOHiYUmW0Hj1Qu9OKAiVQu5Bd547ktVCiMKC1BkB5L3ImeEnCyq3WpRKR6ZTaPgekzqdozwdPs7Lg==",
|
||||
"license": "Apache-2.0",
|
||||
"bin": {
|
||||
"openai": "bin/cli"
|
||||
},
|
||||
"peerDependencies": {
|
||||
"ws": "^8.18.0",
|
||||
"zod": "^3.23.8"
|
||||
},
|
||||
"peerDependenciesMeta": {
|
||||
"ws": {
|
||||
"optional": true
|
||||
},
|
||||
"zod": {
|
||||
"optional": true
|
||||
}
|
||||
}
|
||||
},
|
||||
"node_modules/own-keys": {
|
||||
"version": "1.0.1",
|
||||
"resolved": "https://registry.npmjs.org/own-keys/-/own-keys-1.0.1.tgz",
|
||||
@@ -9827,7 +9859,7 @@
|
||||
"version": "8.18.3",
|
||||
"resolved": "https://registry.npmjs.org/ws/-/ws-8.18.3.tgz",
|
||||
"integrity": "sha512-PEIGCY5tSlUt50cqyMXfCzX+oOPqN0vuGqWzbcJ2xvnkzkq46oOpz7dQaTDBdfICb4N14+GARUDw2XV2N4tvzg==",
|
||||
"dev": true,
|
||||
"devOptional": true,
|
||||
"license": "MIT",
|
||||
"engines": {
|
||||
"node": ">=10.0.0"
|
||||
|
||||
@@ -6,7 +6,7 @@
|
||||
"scripts": {
|
||||
"dev": "concurrently \"npm run dev:frontend\" \"npm run dev:backend\"",
|
||||
"dev:frontend": "webpack serve --mode development --config webpack.config.js",
|
||||
"dev:backend": "nodemon src/index.js",
|
||||
"dev:backend": "nodemon --exitcrash src/index.js",
|
||||
"build": "webpack --config webpack.prod.config.js",
|
||||
"build:prod": "npm run build && npm run start:prod",
|
||||
"start": "npm run build && node src/index.js",
|
||||
@@ -32,6 +32,8 @@
|
||||
"google-auth-library": "^9.0.0",
|
||||
"jsonwebtoken": "^9.0.0",
|
||||
"mssql": "^9.1.0",
|
||||
"nodemailer": "^7.0.5",
|
||||
"openai": "^5.12.0",
|
||||
"react": "^18.2.0",
|
||||
"react-dom": "^18.2.0"
|
||||
},
|
||||
|
||||
@@ -73,10 +73,42 @@ const executeQuery = async (query, params = {}) => {
|
||||
}
|
||||
};
|
||||
|
||||
const executeTransaction = async (callback) => {
|
||||
if (!process.env.DB_SERVER) {
|
||||
throw new Error('Database not configured');
|
||||
}
|
||||
|
||||
let pool;
|
||||
let transaction;
|
||||
|
||||
try {
|
||||
pool = await getPool();
|
||||
transaction = new sql.Transaction(pool);
|
||||
|
||||
await transaction.begin();
|
||||
|
||||
const result = await callback(transaction);
|
||||
|
||||
await transaction.commit();
|
||||
return result;
|
||||
} catch (error) {
|
||||
if (transaction) {
|
||||
try {
|
||||
await transaction.rollback();
|
||||
} catch (rollbackError) {
|
||||
console.error('Transaction rollback failed:', rollbackError);
|
||||
}
|
||||
}
|
||||
console.error('Transaction error:', error);
|
||||
throw error;
|
||||
}
|
||||
};
|
||||
|
||||
module.exports = {
|
||||
config,
|
||||
getPool,
|
||||
testConnection,
|
||||
executeQuery,
|
||||
executeTransaction,
|
||||
sql,
|
||||
};
|
||||
@@ -10,7 +10,7 @@ const dataRoutes = require('./routes/data');
|
||||
const dbConfig = require('./config/database');
|
||||
|
||||
const app = express();
|
||||
const PORT = process.env.PORT || 5000;
|
||||
const PORT = process.env.PORT || 5500;
|
||||
|
||||
// Middleware
|
||||
app.use(cors());
|
||||
|
||||
@@ -76,6 +76,98 @@ router.post('/google', async (req, res) => {
|
||||
}
|
||||
});
|
||||
|
||||
// Google OAuth callback (redirect flow)
|
||||
router.post('/google/callback', async (req, res) => {
|
||||
try {
|
||||
const { code, redirect_uri } = req.body;
|
||||
|
||||
console.log('🔄 Processing OAuth callback with authorization code');
|
||||
|
||||
if (!code) {
|
||||
console.log('❌ No authorization code provided');
|
||||
return res.status(400).json({ error: 'Authorization code is required' });
|
||||
}
|
||||
|
||||
// Exchange authorization code for tokens
|
||||
const tokenResponse = await fetch('https://oauth2.googleapis.com/token', {
|
||||
method: 'POST',
|
||||
headers: {
|
||||
'Content-Type': 'application/x-www-form-urlencoded',
|
||||
},
|
||||
body: new URLSearchParams({
|
||||
client_id: process.env.GOOGLE_CLIENT_ID,
|
||||
client_secret: process.env.GOOGLE_CLIENT_SECRET,
|
||||
code: code,
|
||||
grant_type: 'authorization_code',
|
||||
redirect_uri: redirect_uri,
|
||||
}),
|
||||
});
|
||||
|
||||
if (!tokenResponse.ok) {
|
||||
const errorData = await tokenResponse.text();
|
||||
console.log('❌ Token exchange failed:', errorData);
|
||||
return res.status(400).json({ error: 'Failed to exchange authorization code' });
|
||||
}
|
||||
|
||||
const tokens = await tokenResponse.json();
|
||||
console.log('🎯 Received tokens from Google');
|
||||
|
||||
// Use the ID token to get user info
|
||||
const ticket = await client.verifyIdToken({
|
||||
idToken: tokens.id_token,
|
||||
audience: process.env.GOOGLE_CLIENT_ID,
|
||||
});
|
||||
|
||||
const payload = ticket.getPayload();
|
||||
const googleId = payload['sub'];
|
||||
const email = payload['email'];
|
||||
const name = payload['name'];
|
||||
const picture = payload['picture'];
|
||||
|
||||
console.log(`👤 OAuth callback verified for: ${email}`);
|
||||
|
||||
// Check if email is authorized
|
||||
const authorized = await isEmailAuthorized(email);
|
||||
console.log(`🔒 Email authorization check for ${email}: ${authorized ? 'ALLOWED' : 'DENIED'}`);
|
||||
|
||||
if (!authorized) {
|
||||
console.log(`❌ Access denied for ${email}`);
|
||||
return res.status(403).json({
|
||||
error: 'Access denied',
|
||||
message: 'Your email address is not authorized to access this application'
|
||||
});
|
||||
}
|
||||
|
||||
// Create user object
|
||||
const user = {
|
||||
id: googleId,
|
||||
email,
|
||||
name,
|
||||
picture,
|
||||
google_id: googleId,
|
||||
};
|
||||
console.log('✅ User object created from OAuth callback');
|
||||
|
||||
// Generate JWT token
|
||||
const jwtToken = generateToken(user);
|
||||
|
||||
res.json({
|
||||
success: true,
|
||||
token: jwtToken,
|
||||
user: {
|
||||
id: user.id,
|
||||
email: user.email,
|
||||
name: user.name,
|
||||
picture: user.picture,
|
||||
},
|
||||
});
|
||||
|
||||
} catch (error) {
|
||||
console.error('OAuth callback error:', error);
|
||||
res.status(401).json({ error: 'OAuth authentication failed' });
|
||||
}
|
||||
});
|
||||
|
||||
// Verify JWT token
|
||||
router.get('/verify', authenticateToken, async (req, res) => {
|
||||
try {
|
||||
|
||||
@@ -3,6 +3,39 @@ const { authenticateToken } = require('../../middleware/auth');
|
||||
|
||||
const router = express.Router();
|
||||
|
||||
// Debug: Get JTL Kontierung data for a specific JTL Umsatz (by kZahlungsabgleichUmsatz)
|
||||
router.get('/jtl-kontierung/:jtlId', authenticateToken, async (req, res) => {
|
||||
try {
|
||||
const { executeQuery } = require('../../config/database');
|
||||
const { jtlId } = req.params;
|
||||
|
||||
const query = `
|
||||
SELECT
|
||||
uk.data
|
||||
FROM eazybusiness.dbo.tZahlungsabgleichUmsatz z
|
||||
LEFT JOIN eazybusiness.dbo.tUmsatzKontierung uk
|
||||
ON uk.kZahlungsabgleichUmsatz = z.kZahlungsabgleichUmsatz
|
||||
WHERE z.kZahlungsabgleichUmsatz = @jtlId
|
||||
`;
|
||||
|
||||
const result = await executeQuery(query, { jtlId: parseInt(jtlId, 10) });
|
||||
// Return undefined when no data found (do not lie with empty array/string)
|
||||
if (!result.recordset || result.recordset.length === 0) {
|
||||
return res.json({ data: undefined });
|
||||
}
|
||||
|
||||
// If multiple rows exist, return all; otherwise single object
|
||||
const rows = result.recordset.map(r => ({ data: r.data }));
|
||||
if (rows.length === 1) {
|
||||
return res.json(rows[0]);
|
||||
}
|
||||
return res.json(rows);
|
||||
} catch (error) {
|
||||
console.error('Error fetching JTL Kontierung data:', error);
|
||||
res.status(500).json({ error: 'Failed to fetch JTL Kontierung data' });
|
||||
}
|
||||
});
|
||||
|
||||
// Get accounting items for a specific transaction
|
||||
router.get('/accounting-items/:transactionId', authenticateToken, async (req, res) => {
|
||||
try {
|
||||
|
||||
@@ -71,6 +71,37 @@ router.post('/test-csv-import', async (req, res) => {
|
||||
numericAmount = parseFloat(normalizedAmount) || 0;
|
||||
}
|
||||
|
||||
// Check for existing transaction to prevent duplicates
|
||||
const duplicateCheckQuery = `
|
||||
SELECT COUNT(*) as count FROM fibdash.CSVTransactions
|
||||
WHERE buchungstag = @buchungstag
|
||||
AND wertstellung = @wertstellung
|
||||
AND umsatzart = @umsatzart
|
||||
AND betrag = @betrag
|
||||
AND beguenstigter_zahlungspflichtiger = @beguenstigter_zahlungspflichtiger
|
||||
AND verwendungszweck = @verwendungszweck
|
||||
`;
|
||||
|
||||
const duplicateCheckResult = await executeQuery(duplicateCheckQuery, {
|
||||
buchungstag: transaction['Buchungstag'] || null,
|
||||
wertstellung: transaction['Valutadatum'] || null,
|
||||
umsatzart: transaction['Buchungstext'] || null,
|
||||
betrag: numericAmount,
|
||||
beguenstigter_zahlungspflichtiger: transaction['Beguenstigter/Zahlungspflichtiger'] || null,
|
||||
verwendungszweck: transaction['Verwendungszweck'] || null
|
||||
});
|
||||
|
||||
if (duplicateCheckResult.recordset[0].count > 0) {
|
||||
console.log(`Skipping duplicate transaction at row ${i + 1}: ${transaction['Buchungstag']} - ${numericAmount}`);
|
||||
errors.push({
|
||||
row: i + 1,
|
||||
error: 'Duplicate transaction (already exists in database)',
|
||||
transaction: transaction
|
||||
});
|
||||
errorCount++;
|
||||
continue;
|
||||
}
|
||||
|
||||
const insertQuery = `
|
||||
INSERT INTO fibdash.CSVTransactions
|
||||
(buchungstag, wertstellung, umsatzart, betrag, betrag_original, waehrung,
|
||||
@@ -129,6 +160,7 @@ router.post('/test-csv-import', async (req, res) => {
|
||||
|
||||
// Import CSV transactions to database
|
||||
router.post('/import-csv-transactions', authenticateToken, async (req, res) => {
|
||||
console.log('Importing CSV transactions');
|
||||
try {
|
||||
const { executeQuery } = require('../../config/database');
|
||||
const { transactions, filename, batchId, headers } = req.body;
|
||||
@@ -180,6 +212,7 @@ router.post('/import-csv-transactions', authenticateToken, async (req, res) => {
|
||||
let successCount = 0;
|
||||
let errorCount = 0;
|
||||
const errors = [];
|
||||
console.log('precheck done');
|
||||
|
||||
for (let i = 0; i < transactions.length; i++) {
|
||||
const transaction = transactions[i];
|
||||
@@ -195,9 +228,6 @@ router.post('/import-csv-transactions', authenticateToken, async (req, res) => {
|
||||
validationErrors.push('Betrag is required');
|
||||
}
|
||||
|
||||
if (!transaction['Beguenstigter/Zahlungspflichtiger'] || transaction['Beguenstigter/Zahlungspflichtiger'].trim() === '') {
|
||||
validationErrors.push('Beguenstigter/Zahlungspflichtiger is required');
|
||||
}
|
||||
|
||||
if (validationErrors.length > 2) {
|
||||
console.log('Skipping invalid row ' + (i + 1) + ':', validationErrors);
|
||||
@@ -247,6 +277,37 @@ router.post('/import-csv-transactions', authenticateToken, async (req, res) => {
|
||||
numericAmount = parseFloat(normalizedAmount) || 0;
|
||||
}
|
||||
|
||||
// Check for existing transaction to prevent duplicates
|
||||
const duplicateCheckQuery = `
|
||||
SELECT COUNT(*) as count FROM fibdash.CSVTransactions
|
||||
WHERE buchungstag = @buchungstag
|
||||
AND wertstellung = @wertstellung
|
||||
AND umsatzart = @umsatzart
|
||||
AND betrag = @betrag
|
||||
AND beguenstigter_zahlungspflichtiger = @beguenstigter_zahlungspflichtiger
|
||||
AND verwendungszweck = @verwendungszweck
|
||||
`;
|
||||
|
||||
const duplicateCheckResult = await executeQuery(duplicateCheckQuery, {
|
||||
buchungstag: transaction['Buchungstag'] || null,
|
||||
wertstellung: transaction['Valutadatum'] || null,
|
||||
umsatzart: transaction['Buchungstext'] || null,
|
||||
betrag: numericAmount,
|
||||
beguenstigter_zahlungspflichtiger: transaction['Beguenstigter/Zahlungspflichtiger'] || null,
|
||||
verwendungszweck: transaction['Verwendungszweck'] || null
|
||||
});
|
||||
|
||||
if (duplicateCheckResult.recordset[0].count > 0) {
|
||||
console.log(`Skipping duplicate transaction at row ${i + 1}: ${transaction['Buchungstag']} - ${numericAmount}`);
|
||||
errors.push({
|
||||
row: i + 1,
|
||||
error: 'Duplicate transaction (already exists in database)',
|
||||
transaction: transaction
|
||||
});
|
||||
errorCount++;
|
||||
continue;
|
||||
}
|
||||
|
||||
const insertQuery = `
|
||||
INSERT INTO fibdash.CSVTransactions
|
||||
(buchungstag, wertstellung, umsatzart, betrag, betrag_original, waehrung,
|
||||
@@ -287,6 +348,8 @@ router.post('/import-csv-transactions', authenticateToken, async (req, res) => {
|
||||
errorCount++;
|
||||
}
|
||||
}
|
||||
|
||||
console.log('import done',errors);
|
||||
|
||||
res.json({
|
||||
success: true,
|
||||
@@ -488,7 +551,7 @@ router.post('/import-datev-beleglinks', authenticateToken, async (req, res) => {
|
||||
const updateQuery = `
|
||||
UPDATE eazybusiness.dbo.tUmsatzBeleg
|
||||
SET datevlink = @datevlink
|
||||
WHERE kUmsatzBeleg = @kUmsatzBeleg AND (datevlink IS NULL OR datevlink = '')
|
||||
WHERE kUmsatzBeleg = @kUmsatzBeleg AND (datevlink IS NULL OR datevlink = '' OR datevlink = 'pending')
|
||||
`;
|
||||
|
||||
const updateResult = await executeQuery(updateQuery, {
|
||||
@@ -514,7 +577,7 @@ router.post('/import-datev-beleglinks', authenticateToken, async (req, res) => {
|
||||
const updateQuery = `
|
||||
UPDATE eazybusiness.dbo.tPdfObjekt
|
||||
SET datevlink = @datevlink
|
||||
WHERE kPdfObjekt = @kPdfObjekt AND (datevlink IS NULL OR datevlink = '')
|
||||
WHERE kPdfObjekt = @kPdfObjekt AND (datevlink IS NULL OR datevlink = '' OR datevlink = 'pending')
|
||||
`;
|
||||
|
||||
const updateResult = await executeQuery(updateQuery, {
|
||||
|
||||
@@ -39,14 +39,33 @@ const formatDatevAmount = (amount) => {
|
||||
return Math.abs(amount).toFixed(2).replace('.', ',');
|
||||
};
|
||||
|
||||
const formatDatevDate = (dateString) => {
|
||||
if (!dateString) return '';
|
||||
const parts = dateString.split('.');
|
||||
const formatDatevDate = (date) => {
|
||||
if (!date) return '';
|
||||
|
||||
// Handle Date object
|
||||
if (date instanceof Date) {
|
||||
const day = date.getDate().toString().padStart(2, '0');
|
||||
const month = (date.getMonth() + 1).toString().padStart(2, '0');
|
||||
return day + month;
|
||||
}
|
||||
|
||||
// Handle string date
|
||||
const dateStr = date.toString();
|
||||
const parts = dateStr.split('.');
|
||||
if (parts.length === 3) {
|
||||
const day = parts[0].padStart(2, '0');
|
||||
const month = parts[1].padStart(2, '0');
|
||||
return day + month;
|
||||
}
|
||||
|
||||
// Try to parse as date string
|
||||
const parsedDate = new Date(dateStr);
|
||||
if (!isNaN(parsedDate)) {
|
||||
const day = parsedDate.getDate().toString().padStart(2, '0');
|
||||
const month = (parsedDate.getMonth() + 1).toString().padStart(2, '0');
|
||||
return day + month;
|
||||
}
|
||||
|
||||
return '';
|
||||
};
|
||||
|
||||
@@ -55,13 +74,219 @@ const quote = (str, maxLen = 60) => {
|
||||
return '"' + str.slice(0, maxLen).replace(/"/g, '""') + '"';
|
||||
};
|
||||
|
||||
// Parse konto field which might contain multiple accounts like "5400+5300"
|
||||
const parseKonto = (konto) => {
|
||||
if (!konto) return '';
|
||||
// Take the first account number if multiple are present
|
||||
const parts = konto.split('+');
|
||||
return parts[0].trim();
|
||||
};
|
||||
|
||||
// DATEV export endpoint
|
||||
router.get('/datev/:timeRange', authenticateToken, async (req, res) => {
|
||||
try {
|
||||
const { timeRange } = req.params;
|
||||
// TODO: Update to use database queries instead of CSV file
|
||||
res.status(501).json({ error: 'DATEV export temporarily disabled - use database-based queries' });
|
||||
return;
|
||||
const { executeQuery } = require('../../config/database');
|
||||
|
||||
// Parse the time range to get start and end dates
|
||||
let startDate, endDate;
|
||||
|
||||
if (timeRange.includes('-Q')) {
|
||||
// Quarter format: 2025-Q1
|
||||
const [year, quarterPart] = timeRange.split('-Q');
|
||||
const quarter = parseInt(quarterPart, 10);
|
||||
const startMonth = (quarter - 1) * 3 + 1;
|
||||
const endMonth = startMonth + 2;
|
||||
|
||||
startDate = new Date(year, startMonth - 1, 1);
|
||||
endDate = new Date(year, endMonth - 1, new Date(year, endMonth, 0).getDate());
|
||||
} else if (timeRange.length === 4) {
|
||||
// Year format: 2025
|
||||
startDate = new Date(timeRange, 0, 1);
|
||||
endDate = new Date(timeRange, 11, 31);
|
||||
} else {
|
||||
// Month format: 2025-03
|
||||
const [year, month] = timeRange.split('-');
|
||||
startDate = new Date(year, parseInt(month) - 1, 1);
|
||||
endDate = new Date(year, parseInt(month), 0);
|
||||
}
|
||||
|
||||
// Format dates for SQL query
|
||||
const sqlStartDate = startDate.toISOString().split('T')[0];
|
||||
const sqlEndDate = endDate.toISOString().split('T')[0];
|
||||
|
||||
// Query to get all DATEV data with proper joins
|
||||
// This handles multiple documents per transaction by creating separate rows
|
||||
const query = `
|
||||
WITH DatevDocuments AS (
|
||||
-- Get documents from tUmsatzBeleg
|
||||
SELECT
|
||||
uk.kZahlungsabgleichUmsatz,
|
||||
zu.fBetrag as umsatz_brutto,
|
||||
CASE WHEN zu.fBetrag < 0 THEN 'H' ELSE 'S' END as soll_haben_kz,
|
||||
JSON_VALUE(uk.data, '$.konto1') as konto,
|
||||
'' as gegenkonto, -- No creditorID in tUmsatzBeleg
|
||||
-- BU determination based on amount and konto type
|
||||
CASE
|
||||
WHEN JSON_VALUE(uk.data, '$.konto1') IN ('3720', '3740', '2100', '1460', '1462') THEN ''
|
||||
WHEN zu.fBetrag > 0 THEN ''
|
||||
WHEN JSON_VALUE(uk.data, '$.konto1') LIKE '5%' THEN '9' -- 19% for purchases
|
||||
WHEN JSON_VALUE(uk.data, '$.konto1') LIKE '6%' THEN '9' -- 19% for expenses
|
||||
ELSE ''
|
||||
END as bu,
|
||||
FORMAT(zu.dBuchungsdatum, 'Mdd') as buchungsdatum_mdd,
|
||||
zu.dBuchungsdatum,
|
||||
'' as rechnungsnummer, -- No invoice number in tUmsatzBeleg
|
||||
zu.cVerwendungszweck as buchungstext,
|
||||
ub.datevlink as beleglink,
|
||||
1 as priority -- tUmsatzBeleg has priority
|
||||
FROM tUmsatzKontierung uk
|
||||
INNER JOIN tZahlungsabgleichUmsatz zu ON uk.kZahlungsabgleichUmsatz = zu.kZahlungsabgleichUmsatz
|
||||
INNER JOIN tUmsatzBeleg ub ON ub.kZahlungsabgleichUmsatz = zu.kZahlungsabgleichUmsatz
|
||||
WHERE ub.datevlink IS NOT NULL
|
||||
AND zu.dBuchungsdatum >= @startDate
|
||||
AND zu.dBuchungsdatum <= @endDate
|
||||
|
||||
UNION ALL
|
||||
|
||||
-- Get documents from tPdfObjekt via tZahlungsabgleichUmsatzLink
|
||||
SELECT
|
||||
uk.kZahlungsabgleichUmsatz,
|
||||
zu.fBetrag as umsatz_brutto,
|
||||
CASE WHEN zu.fBetrag < 0 THEN 'H' ELSE 'S' END as soll_haben_kz,
|
||||
JSON_VALUE(uk.data, '$.konto1') as konto,
|
||||
COALESCE(JSON_VALUE(po.extraction, '$.creditorID'), '') as gegenkonto,
|
||||
-- BU determination based on amount and konto type
|
||||
CASE
|
||||
WHEN JSON_VALUE(uk.data, '$.konto1') IN ('3720', '3740', '2100', '1460', '1462') THEN ''
|
||||
WHEN zu.fBetrag > 0 THEN ''
|
||||
WHEN JSON_VALUE(uk.data, '$.konto1') LIKE '5%' THEN '9' -- 19% for purchases
|
||||
WHEN JSON_VALUE(uk.data, '$.konto1') LIKE '6%' THEN '9' -- 19% for expenses
|
||||
ELSE ''
|
||||
END as bu,
|
||||
FORMAT(zu.dBuchungsdatum, 'Mdd') as buchungsdatum_mdd,
|
||||
zu.dBuchungsdatum,
|
||||
COALESCE(JSON_VALUE(po.extraction, '$.invoice_number'), '') as rechnungsnummer,
|
||||
zu.cVerwendungszweck as buchungstext,
|
||||
po.datevlink as beleglink,
|
||||
2 as priority -- tPdfObjekt has lower priority
|
||||
FROM tUmsatzKontierung uk
|
||||
INNER JOIN tZahlungsabgleichUmsatz zu ON uk.kZahlungsabgleichUmsatz = zu.kZahlungsabgleichUmsatz
|
||||
INNER JOIN tZahlungsabgleichUmsatzLink zul ON zu.kZahlungsabgleichUmsatz = zul.kZahlungsabgleichUmsatz
|
||||
AND zul.linktype = 'kLieferantenBestellung'
|
||||
INNER JOIN tPdfObjekt po ON zul.linktarget = po.kLieferantenbestellung
|
||||
WHERE po.datevlink IS NOT NULL
|
||||
AND zu.dBuchungsdatum >= @startDate
|
||||
AND zu.dBuchungsdatum <= @endDate
|
||||
|
||||
UNION ALL
|
||||
|
||||
-- Get transactions without documents
|
||||
SELECT
|
||||
uk.kZahlungsabgleichUmsatz,
|
||||
zu.fBetrag as umsatz_brutto,
|
||||
CASE WHEN zu.fBetrag < 0 THEN 'H' ELSE 'S' END as soll_haben_kz,
|
||||
JSON_VALUE(uk.data, '$.konto1') as konto,
|
||||
'' as gegenkonto,
|
||||
-- BU determination based on amount and konto type
|
||||
CASE
|
||||
WHEN JSON_VALUE(uk.data, '$.konto1') IN ('3720', '3740', '2100', '1460', '1462') THEN ''
|
||||
WHEN zu.fBetrag > 0 THEN ''
|
||||
WHEN JSON_VALUE(uk.data, '$.konto1') LIKE '5%' THEN '9' -- 19% for purchases
|
||||
WHEN JSON_VALUE(uk.data, '$.konto1') LIKE '6%' THEN '9' -- 19% for expenses
|
||||
ELSE ''
|
||||
END as bu,
|
||||
FORMAT(zu.dBuchungsdatum, 'Mdd') as buchungsdatum_mdd,
|
||||
zu.dBuchungsdatum,
|
||||
'' as rechnungsnummer,
|
||||
zu.cVerwendungszweck as buchungstext,
|
||||
'' as beleglink,
|
||||
3 as priority -- No documents has lowest priority
|
||||
FROM tUmsatzKontierung uk
|
||||
INNER JOIN tZahlungsabgleichUmsatz zu ON uk.kZahlungsabgleichUmsatz = zu.kZahlungsabgleichUmsatz
|
||||
WHERE zu.dBuchungsdatum >= @startDate
|
||||
AND zu.dBuchungsdatum <= @endDate
|
||||
AND NOT EXISTS (
|
||||
SELECT 1 FROM tUmsatzBeleg ub2
|
||||
WHERE ub2.kZahlungsabgleichUmsatz = zu.kZahlungsabgleichUmsatz
|
||||
AND ub2.datevlink IS NOT NULL
|
||||
)
|
||||
AND NOT EXISTS (
|
||||
SELECT 1 FROM tZahlungsabgleichUmsatzLink zul2
|
||||
INNER JOIN tPdfObjekt po2 ON zul2.linktarget = po2.kLieferantenbestellung
|
||||
WHERE zul2.kZahlungsabgleichUmsatz = zu.kZahlungsabgleichUmsatz
|
||||
AND zul2.linktype = 'kLieferantenBestellung'
|
||||
AND po2.datevlink IS NOT NULL
|
||||
)
|
||||
)
|
||||
SELECT
|
||||
*,
|
||||
ROW_NUMBER() OVER (PARTITION BY kZahlungsabgleichUmsatz, beleglink ORDER BY priority) as rn
|
||||
FROM DatevDocuments
|
||||
ORDER BY dBuchungsdatum DESC, kZahlungsabgleichUmsatz, priority
|
||||
`;
|
||||
|
||||
const result = await executeQuery(query, {
|
||||
startDate: sqlStartDate,
|
||||
endDate: sqlEndDate
|
||||
});
|
||||
|
||||
// Format data for DATEV CSV
|
||||
const datevRows = [];
|
||||
|
||||
// Build header
|
||||
const periodStart = startDate.getFullYear() +
|
||||
('0' + (startDate.getMonth() + 1)).slice(-2) +
|
||||
('0' + startDate.getDate()).slice(-2);
|
||||
const periodEnd = endDate.getFullYear() +
|
||||
('0' + (endDate.getMonth() + 1)).slice(-2) +
|
||||
('0' + endDate.getDate()).slice(-2);
|
||||
|
||||
datevRows.push(buildDatevHeader(periodStart, periodEnd));
|
||||
datevRows.push(DATEV_COLS);
|
||||
|
||||
// Process each transaction
|
||||
result.recordset.forEach(row => {
|
||||
// Skip duplicate rows (keep only the first occurrence of each transaction+beleglink combination)
|
||||
if (row.rn > 1) return;
|
||||
|
||||
const datevRow = [
|
||||
formatDatevAmount(row.umsatz_brutto), // Umsatz (ohne Soll/Haben-Kz)
|
||||
row.soll_haben_kz, // Soll/Haben-Kennzeichen
|
||||
'', // WKZ Umsatz
|
||||
'', // Kurs
|
||||
'', // Basis-Umsatz
|
||||
'', // WKZ Basis-Umsatz
|
||||
parseKonto(row.konto), // Konto (parsed)
|
||||
row.gegenkonto || '', // Gegenkonto (ohne BU-Schlüssel)
|
||||
row.bu || '', // BU-Schlüssel
|
||||
row.buchungsdatum_mdd || '', // Belegdatum (MDD format)
|
||||
quote(row.rechnungsnummer || ''), // Belegfeld 1 (invoice number)
|
||||
'', // Belegfeld 2
|
||||
'', // Skonto
|
||||
quote(row.buchungstext || ''), // Buchungstext
|
||||
'', // Postensperre
|
||||
'', // Diverse Adressnummer
|
||||
'', // Geschäftspartnerbank
|
||||
'', // Sachverhalt
|
||||
'', // Zinssperre
|
||||
row.beleglink || '' // Beleglink
|
||||
].join(';');
|
||||
|
||||
datevRows.push(datevRow);
|
||||
});
|
||||
|
||||
// Generate CSV content
|
||||
const csvContent = datevRows.join('\n');
|
||||
|
||||
// Set headers for CSV download
|
||||
const filename = `EXTF_${timeRange.replace('-', '_')}.csv`;
|
||||
res.setHeader('Content-Type', 'text/csv; charset=windows-1252');
|
||||
res.setHeader('Content-Disposition', `attachment; filename="${filename}"`);
|
||||
|
||||
// Send CSV content
|
||||
res.send(csvContent);
|
||||
|
||||
} catch (error) {
|
||||
console.error('Error generating DATEV export:', error);
|
||||
res.status(500).json({ error: 'Failed to generate DATEV export' });
|
||||
|
||||
416
src/routes/data/documentProcessing.js
Normal file
416
src/routes/data/documentProcessing.js
Normal file
@@ -0,0 +1,416 @@
|
||||
const express = require('express');
|
||||
const { authenticateToken } = require('../../middleware/auth');
|
||||
const { executeQuery, executeTransaction } = require('../../config/database');
|
||||
const sql = require('mssql');
|
||||
const nodemailer = require('nodemailer');
|
||||
|
||||
const router = express.Router();
|
||||
|
||||
// Get document processing status
|
||||
router.get('/document-status', authenticateToken, async (req, res) => {
|
||||
try {
|
||||
console.log('Document status endpoint called');
|
||||
const queries = {
|
||||
needMarkdownUmsatz: "SELECT COUNT(*) as count FROM tUmsatzBeleg WHERE markDown is null",
|
||||
needMarkdownPdf: "SELECT COUNT(*) as count FROM tPdfObjekt WHERE markDown is null",
|
||||
needExtractionUmsatz: "SELECT COUNT(*) as count FROM tUmsatzBeleg WHERE markDown is not null and extraction is null",
|
||||
needExtractionPdf: "SELECT COUNT(*) as count FROM tPdfObjekt WHERE markDown is not null and extraction is null",
|
||||
needDatevSyncUmsatz: "SELECT COUNT(*) as count FROM tUmsatzBeleg WHERE markDown is not null and datevlink is null",
|
||||
needDatevSyncPdf: "SELECT COUNT(*) as count FROM tPdfObjekt WHERE markDown is not null and datevlink is null",
|
||||
needDatevUploadUmsatz: "SELECT COUNT(*) as count FROM tUmsatzBeleg WHERE datevlink = 'pending'",
|
||||
needDatevUploadPdf: "SELECT COUNT(*) as count FROM tPdfObjekt WHERE datevlink = 'pending'"
|
||||
};
|
||||
|
||||
const results = {};
|
||||
for (const [key, query] of Object.entries(queries)) {
|
||||
const result = await executeQuery(query);
|
||||
results[key] = result.recordset[0].count;
|
||||
}
|
||||
|
||||
const status = {
|
||||
needMarkdown: results.needMarkdownUmsatz + results.needMarkdownPdf,
|
||||
needExtraction: results.needExtractionUmsatz + results.needExtractionPdf,
|
||||
needDatevSync: results.needDatevSyncUmsatz + results.needDatevSyncPdf,
|
||||
needDatevUpload: results.needDatevUploadUmsatz + results.needDatevUploadPdf,
|
||||
details: {
|
||||
markdown: {
|
||||
umsatzBeleg: results.needMarkdownUmsatz,
|
||||
pdfObjekt: results.needMarkdownPdf
|
||||
},
|
||||
extraction: {
|
||||
umsatzBeleg: results.needExtractionUmsatz,
|
||||
pdfObjekt: results.needExtractionPdf
|
||||
},
|
||||
datevSync: {
|
||||
umsatzBeleg: results.needDatevSyncUmsatz,
|
||||
pdfObjekt: results.needDatevSyncPdf
|
||||
},
|
||||
datevUpload: {
|
||||
umsatzBeleg: results.needDatevUploadUmsatz,
|
||||
pdfObjekt: results.needDatevUploadPdf
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
console.log('Document status computed:', status);
|
||||
res.json(status);
|
||||
} catch (error) {
|
||||
console.error('Error fetching document processing status:', error);
|
||||
res.status(500).json({ error: 'Failed to fetch document processing status' });
|
||||
}
|
||||
});
|
||||
|
||||
// Process markdown conversion
|
||||
router.post('/process-markdown', authenticateToken, async (req, res) => {
|
||||
try {
|
||||
const { OpenAI } = require('openai');
|
||||
|
||||
// Check environment for OpenAI API key
|
||||
if (!process.env.OPENAI_API_KEY) {
|
||||
return res.status(500).json({ error: 'OpenAI API key not configured' });
|
||||
}
|
||||
|
||||
const openai = new OpenAI({
|
||||
apiKey: process.env.OPENAI_API_KEY,
|
||||
});
|
||||
|
||||
await executeTransaction(async (transaction) => {
|
||||
// Process UmsatzBeleg documents
|
||||
const umsatzResult = await new sql.Request(transaction).query(
|
||||
"SELECT TOP 1 kUmsatzBeleg, content FROM tUmsatzBeleg WHERE markDown is null"
|
||||
);
|
||||
|
||||
if (umsatzResult.recordset.length > 0) {
|
||||
const { kUmsatzBeleg, content } = umsatzResult.recordset[0];
|
||||
|
||||
const response = await openai.responses.create({
|
||||
model: "gpt-4o",
|
||||
input: [
|
||||
{ "role": "developer", "content": [{ "type": "input_text", "text": "Convert to Markdown" }] },
|
||||
{ "role": "user", "content": [{ "type": "input_file", "filename": "invoice.pdf", "file_data": "data:application/pdf;base64," + content.toString('base64') }] }
|
||||
],
|
||||
text: {
|
||||
"format": {
|
||||
"type": "json_schema", "name": "markdown", "strict": true, "schema": { "type": "object", "properties": {
|
||||
"output": { "type": "string", "description": "Input converted to Markdown" }
|
||||
}, "required": ["output"], "additionalProperties": false }
|
||||
}
|
||||
},
|
||||
tools: [],
|
||||
store: false
|
||||
});
|
||||
|
||||
const markdown = JSON.parse(response.output_text);
|
||||
|
||||
await new sql.Request(transaction)
|
||||
.input('kUmsatzBeleg', kUmsatzBeleg)
|
||||
.input('markDown', markdown.output)
|
||||
.query("UPDATE tUmsatzBeleg SET markDown = @markDown WHERE kUmsatzBeleg = @kUmsatzBeleg");
|
||||
}
|
||||
|
||||
// Process PdfObjekt documents
|
||||
const pdfResult = await new sql.Request(transaction).query(
|
||||
"SELECT TOP 1 kPdfObjekt, content FROM tPdfObjekt WHERE markDown is null"
|
||||
);
|
||||
|
||||
if (pdfResult.recordset.length > 0) {
|
||||
const { kPdfObjekt, content } = pdfResult.recordset[0];
|
||||
|
||||
const response = await openai.responses.create({
|
||||
model: "gpt-4o",
|
||||
input: [
|
||||
{ "role": "developer", "content": [{ "type": "input_text", "text": "Convert to Markdown" }] },
|
||||
{ "role": "user", "content": [{ "type": "input_file", "filename": "invoice.pdf", "file_data": "data:application/pdf;base64," + content.toString('base64') }] }
|
||||
],
|
||||
text: {
|
||||
"format": {
|
||||
"type": "json_schema", "name": "markdown", "strict": true, "schema": { "type": "object", "properties": {
|
||||
"output": { "type": "string", "description": "Input converted to Markdown" }
|
||||
}, "required": ["output"], "additionalProperties": false }
|
||||
}
|
||||
},
|
||||
tools: [],
|
||||
store: false
|
||||
});
|
||||
|
||||
const markdown = JSON.parse(response.output_text);
|
||||
|
||||
await new sql.Request(transaction)
|
||||
.input('kPdfObjekt', kPdfObjekt)
|
||||
.input('markDown', markdown.output)
|
||||
.query("UPDATE tPdfObjekt SET markDown = @markDown WHERE kPdfObjekt = @kPdfObjekt");
|
||||
}
|
||||
});
|
||||
|
||||
res.json({ success: true, message: 'Markdown processing completed' });
|
||||
} catch (error) {
|
||||
console.error('Error processing markdown:', error);
|
||||
res.status(500).json({ error: 'Failed to process markdown: ' + error.message });
|
||||
}
|
||||
});
|
||||
|
||||
// Process data extraction
|
||||
router.post('/process-extraction', authenticateToken, async (req, res) => {
|
||||
try {
|
||||
const { OpenAI } = require('openai');
|
||||
|
||||
if (!process.env.OPENAI_API_KEY) {
|
||||
return res.status(500).json({ error: 'OpenAI API key not configured' });
|
||||
}
|
||||
|
||||
const openai = new OpenAI({
|
||||
apiKey: process.env.OPENAI_API_KEY,
|
||||
});
|
||||
|
||||
await executeTransaction(async (transaction) => {
|
||||
// Get creditor IDs for extraction
|
||||
const creditorResult = await new sql.Request(transaction).query(
|
||||
"SELECT kreditorId FROM fibdash.Kreditor ORDER BY kreditorId"
|
||||
);
|
||||
const creditorIDs = creditorResult.recordset.map(r => r.kreditorId).join(', ');
|
||||
|
||||
// Process UmsatzBeleg documents
|
||||
const umsatzResult = await new sql.Request(transaction).query(
|
||||
"SELECT TOP 1 kUmsatzBeleg, markDown FROM tUmsatzBeleg WHERE markDown is not null and extraction is null"
|
||||
);
|
||||
|
||||
if (umsatzResult.recordset.length > 0) {
|
||||
const { kUmsatzBeleg, markDown } = umsatzResult.recordset[0];
|
||||
|
||||
const response = await openai.responses.create({
|
||||
model: "gpt-5-mini",
|
||||
input: [
|
||||
{ "role": "developer", "content": [{ "type": "input_text", "text": `Extract specified information from provided input and structure it in a JSON format.
|
||||
|
||||
The aim is to accurately identify and capture the following elements:
|
||||
- Rechnungsdatum/Belegdatum (Invoice Date/Document Date),
|
||||
- Rechnungsnummer/Belegnummer (Invoice Number/Document Number),
|
||||
- Netto Betrag (Net Amount),
|
||||
- Brutto Betrag (Gross Amount),
|
||||
- and Absender (Sender).
|
||||
|
||||
# Steps
|
||||
|
||||
1. **Identify Dates**: Find and extract the invoice or document date (Rechnungsdatum/Belegdatum) from the input text.
|
||||
2. **Extract Numbers**: Locate and pull out the invoice or document number (Rechnungsnummer/Belegnummer).
|
||||
3. **Determine Amounts**: Identify the net amount (Netto Betrag) and the gross amount (Brutto Betrag) and the currency in the text.
|
||||
4. **Source the Sender**: Extract the sender's information (Absender, Country).
|
||||
5. **Structure Data**: Organize the extracted information into a JSON format following the specified schema.
|
||||
|
||||
# Notes
|
||||
|
||||
- Ensure that dates are formatted consistently.
|
||||
- Be mindful of various numerical representations (e.g., with commas or periods).
|
||||
- The sender's information might include company names, so recognize various formats.
|
||||
- Prioritize accuracy in identifying the correct fields, as there can be similar text elements present.
|
||||
|
||||
Also select the CreditorID, from that List: ${creditorIDs}` }] },
|
||||
{ "role": "user", "content": [{ "type": "input_text", "text": markDown }] }
|
||||
],
|
||||
text: {
|
||||
"format": {
|
||||
"type": "json_schema", "name": "invoice", "strict": true, "schema": { "type": "object", "properties": {
|
||||
"date": { "type": "string", "description": "Rechungsdatum / Belegdatum in ISO 8601" },
|
||||
"invoice_number": { "type": "string", "description": "Rechnungsnummer / Belegnummer / Invoicenr" },
|
||||
"net_amounts_and_tax": {
|
||||
"type": "array", "description": "Liste von Nettobeträgen mit jeweiligem Steuersatz und Steuerbetrag, ein Listeneintrag pro Steuersatz",
|
||||
"items": { "type": "object", "properties": {
|
||||
"net_amount": { "type": "number", "description": "Netto Betrag" },
|
||||
"tax_rate": { "type": "number", "description": "Steuersatz in Prozent" },
|
||||
"tax_amount": { "type": "number", "description": "Steuerbetrag" }
|
||||
}, "required": ["net_amount", "tax_rate", "tax_amount"], "additionalProperties": false }
|
||||
},
|
||||
"gross_amount": { "type": "number", "description": "Brutto Betrag (muss der Summe aller net_amount + tax_amount entsprechen)" },
|
||||
"currency": { "type": "string", "description": "currency code in ISO 4217" },
|
||||
"country": { "type": "string", "description": "country of origin in ISO 3166" },
|
||||
"sender": { "type": "string", "description": "Absender" },
|
||||
"creditorID": { "type": "string", "description": "CreditorID or empty if unknown" }
|
||||
}, "required": ["date", "invoice_number", "net_amounts_and_tax", "gross_amount", "currency", "country", "sender", "creditorID"], "additionalProperties": false }
|
||||
}
|
||||
},
|
||||
tools: [],
|
||||
store: false
|
||||
});
|
||||
|
||||
const extraction = JSON.parse(response.output_text);
|
||||
|
||||
await new sql.Request(transaction)
|
||||
.input('kUmsatzBeleg', kUmsatzBeleg)
|
||||
.input('extraction', JSON.stringify(extraction))
|
||||
.query("UPDATE tUmsatzBeleg SET extraction = @extraction WHERE kUmsatzBeleg = @kUmsatzBeleg");
|
||||
}
|
||||
|
||||
// Process PdfObjekt documents
|
||||
const pdfResult = await new sql.Request(transaction).query(
|
||||
"SELECT TOP 1 kPdfObjekt, markDown FROM tPdfObjekt WHERE markDown is not null and extraction is null"
|
||||
);
|
||||
|
||||
if (pdfResult.recordset.length > 0) {
|
||||
const { kPdfObjekt, markDown } = pdfResult.recordset[0];
|
||||
|
||||
const response = await openai.responses.create({
|
||||
model: "gpt-5-mini",
|
||||
input: [
|
||||
{ "role": "developer", "content": [{ "type": "input_text", "text": `Extract specified information from provided input and structure it in a JSON format.
|
||||
|
||||
The aim is to accurately identify and capture the following elements:
|
||||
- Rechnungsdatum/Belegdatum (Invoice Date/Document Date),
|
||||
- Rechnungsnummer/Belegnummer (Invoice Number/Document Number),
|
||||
- Netto Betrag (Net Amount),
|
||||
- Brutto Betrag (Gross Amount),
|
||||
- and Absender (Sender).
|
||||
|
||||
# Steps
|
||||
|
||||
1. **Identify Dates**: Find and extract the invoice or document date (Rechnungsdatum/Belegdatum) from the input text.
|
||||
2. **Extract Numbers**: Locate and pull out the invoice or document number (Rechnungsnummer/Belegnummer).
|
||||
3. **Determine Amounts**: Identify the net amount (Netto Betrag) and the gross amount (Brutto Betrag) and the currency in the text.
|
||||
4. **Source the Sender**: Extract the sender's information (Absender, Country).
|
||||
5. **Structure Data**: Organize the extracted information into a JSON format following the specified schema.
|
||||
|
||||
# Notes
|
||||
|
||||
- Ensure that dates are formatted consistently.
|
||||
- Be mindful of various numerical representations (e.g., with commas or periods).
|
||||
- The sender's information might include company names, so recognize various formats.
|
||||
- Prioritize accuracy in identifying the correct fields, as there can be similar text elements present.
|
||||
|
||||
Also select the CreditorID, from that List: ${creditorIDs}` }] },
|
||||
{ "role": "user", "content": [{ "type": "input_text", "text": markDown }] }
|
||||
],
|
||||
text: {
|
||||
"format": {
|
||||
"type": "json_schema", "name": "invoice", "strict": true, "schema": { "type": "object", "properties": {
|
||||
"date": { "type": "string", "description": "Rechungsdatum / Belegdatum in ISO 8601" },
|
||||
"invoice_number": { "type": "string", "description": "Rechnungsnummer / Belegnummer / Invoicenr" },
|
||||
"net_amounts_and_tax": {
|
||||
"type": "array", "description": "Liste von Nettobeträgen mit jeweiligem Steuersatz und Steuerbetrag, ein Listeneintrag pro Steuersatz",
|
||||
"items": { "type": "object", "properties": {
|
||||
"net_amount": { "type": "number", "description": "Netto Betrag" },
|
||||
"tax_rate": { "type": "number", "description": "Steuersatz in Prozent" },
|
||||
"tax_amount": { "type": "number", "description": "Steuerbetrag" }
|
||||
}, "required": ["net_amount", "tax_rate", "tax_amount"], "additionalProperties": false }
|
||||
},
|
||||
"gross_amount": { "type": "number", "description": "Brutto Betrag (muss der Summe aller net_amount + tax_amount entsprechen)" },
|
||||
"currency": { "type": "string", "description": "currency code in ISO 4217" },
|
||||
"country": { "type": "string", "description": "country of origin in ISO 3166" },
|
||||
"sender": { "type": "string", "description": "Absender" },
|
||||
"creditorID": { "type": "string", "description": "CreditorID or empty if unknown" }
|
||||
}, "required": ["date", "invoice_number", "net_amounts_and_tax", "gross_amount", "currency", "country", "sender", "creditorID"], "additionalProperties": false }
|
||||
}
|
||||
},
|
||||
tools: [],
|
||||
store: false
|
||||
});
|
||||
|
||||
const extraction = JSON.parse(response.output_text);
|
||||
|
||||
await new sql.Request(transaction)
|
||||
.input('kPdfObjekt', kPdfObjekt)
|
||||
.input('extraction', JSON.stringify(extraction))
|
||||
.query("UPDATE tPdfObjekt SET extraction = @extraction WHERE kPdfObjekt = @kPdfObjekt");
|
||||
}
|
||||
});
|
||||
|
||||
res.json({ success: true, message: 'Extraction processing completed' });
|
||||
} catch (error) {
|
||||
console.error('Error processing extraction:', error);
|
||||
res.status(500).json({ error: 'Failed to process extraction: ' + error.message });
|
||||
}
|
||||
});
|
||||
|
||||
// Process Datev sync
|
||||
router.post('/process-datev-sync', authenticateToken, async (req, res) => {
|
||||
try {
|
||||
const transporter = nodemailer.createTransport({
|
||||
host: "smtp.gmail.com",
|
||||
port: 587,
|
||||
secure: false, // true for 465, false for other ports
|
||||
auth: {
|
||||
user: "sebgreenbus@gmail.com",
|
||||
pass: "abrp idub thbi kdws", // For Gmail, you might need an app-specific password
|
||||
},
|
||||
});
|
||||
|
||||
await executeTransaction(async (transaction) => {
|
||||
// Process UmsatzBeleg documents
|
||||
const umsatzResult = await new sql.Request(transaction).query(
|
||||
"SELECT TOP 1 kUmsatzBeleg, content FROM tUmsatzBeleg WHERE markDown is not null and datevlink is null"
|
||||
);
|
||||
|
||||
if (umsatzResult.recordset.length > 0) {
|
||||
const { kUmsatzBeleg, content } = umsatzResult.recordset[0];
|
||||
|
||||
const mailOptions = {
|
||||
from: '"Growheads" <sebgreenbus@gmail.com>',
|
||||
to: "97bfd9eb-770f-481a-accb-e69649d36a9e@uploadmail.datev.de",
|
||||
subject: `Beleg ${kUmsatzBeleg} für Datev`,
|
||||
text: "", // No body text as requested
|
||||
attachments: [
|
||||
{
|
||||
filename: `UmsatzBeleg${kUmsatzBeleg}.pdf`,
|
||||
content: content,
|
||||
contentType: "application/pdf",
|
||||
},
|
||||
],
|
||||
};
|
||||
|
||||
try {
|
||||
let info = await transporter.sendMail(mailOptions);
|
||||
console.log("Message sent: %s", info.messageId);
|
||||
|
||||
await new sql.Request(transaction)
|
||||
.input('kUmsatzBeleg', kUmsatzBeleg)
|
||||
.input('datevlink', 'pending')
|
||||
.query("UPDATE tUmsatzBeleg SET datevlink = @datevlink WHERE kUmsatzBeleg = @kUmsatzBeleg");
|
||||
} catch (emailError) {
|
||||
console.error("Error sending email:", emailError);
|
||||
throw emailError;
|
||||
}
|
||||
}
|
||||
|
||||
// Process PdfObjekt documents
|
||||
const pdfResult = await new sql.Request(transaction).query(
|
||||
"SELECT TOP 1 kPdfObjekt, content FROM tPdfObjekt WHERE markDown is not null and datevlink is null"
|
||||
);
|
||||
|
||||
if (pdfResult.recordset.length > 0) {
|
||||
const { kPdfObjekt, content } = pdfResult.recordset[0];
|
||||
|
||||
const mailOptions = {
|
||||
from: '"Growheads" <sebgreenbus@gmail.com>',
|
||||
to: "97bfd9eb-770f-481a-accb-e69649d36a9e@uploadmail.datev.de",
|
||||
subject: `Rechnung ${kPdfObjekt} für Datev`,
|
||||
text: "", // No body text as requested
|
||||
attachments: [
|
||||
{
|
||||
filename: `Rechnung${kPdfObjekt}.pdf`,
|
||||
content: content,
|
||||
contentType: "application/pdf",
|
||||
},
|
||||
],
|
||||
};
|
||||
|
||||
try {
|
||||
let info = await transporter.sendMail(mailOptions);
|
||||
console.log("Message sent: %s", info.messageId);
|
||||
|
||||
await new sql.Request(transaction)
|
||||
.input('kPdfObjekt', kPdfObjekt)
|
||||
.input('datevlink', 'pending')
|
||||
.query("UPDATE tPdfObjekt SET datevlink = @datevlink WHERE kPdfObjekt = @kPdfObjekt");
|
||||
} catch (emailError) {
|
||||
console.error("Error sending email:", emailError);
|
||||
throw emailError;
|
||||
}
|
||||
}
|
||||
});
|
||||
|
||||
res.json({ success: true, message: 'Datev sync processing completed' });
|
||||
} catch (error) {
|
||||
console.error('Error processing Datev sync:', error);
|
||||
res.status(500).json({ error: 'Failed to process Datev sync: ' + error.message });
|
||||
}
|
||||
});
|
||||
|
||||
module.exports = router;
|
||||
@@ -8,6 +8,7 @@ const kreditors = require('./kreditors');
|
||||
const bankingTransactions = require('./bankingTransactions');
|
||||
const accountingItems = require('./accountingItems');
|
||||
const csvImport = require('./csvImport');
|
||||
const documentProcessing = require('./documentProcessing');
|
||||
|
||||
const router = express.Router();
|
||||
|
||||
@@ -20,5 +21,6 @@ router.use(kreditors);
|
||||
router.use(bankingTransactions);
|
||||
router.use(accountingItems);
|
||||
router.use(csvImport);
|
||||
router.use(documentProcessing);
|
||||
|
||||
module.exports = router;
|
||||
@@ -10,6 +10,25 @@ router.get('/transactions/:timeRange', authenticateToken, async (req, res) => {
|
||||
const { timeRange } = req.params;
|
||||
|
||||
const { executeQuery } = require('../../config/database');
|
||||
|
||||
// Build WHERE clause based on timeRange format
|
||||
let timeWhereClause = '';
|
||||
if (timeRange.includes('-Q')) {
|
||||
// Quarter format: 2025-Q2
|
||||
const [year, quarterPart] = timeRange.split('-Q');
|
||||
const quarter = parseInt(quarterPart, 10);
|
||||
const startMonth = (quarter - 1) * 3 + 1;
|
||||
const endMonth = startMonth + 2;
|
||||
timeWhereClause = `WHERE YEAR(csv.parsed_date) = ${year} AND MONTH(csv.parsed_date) BETWEEN ${startMonth} AND ${endMonth}`;
|
||||
} else if (timeRange.length === 4) {
|
||||
// Year format: 2025
|
||||
timeWhereClause = `WHERE YEAR(csv.parsed_date) = ${timeRange}`;
|
||||
} else {
|
||||
// Month format: 2025-07
|
||||
const [year, month] = timeRange.split('-');
|
||||
timeWhereClause = `WHERE YEAR(csv.parsed_date) = ${year} AND MONTH(csv.parsed_date) = ${parseInt(month, 10)}`;
|
||||
}
|
||||
|
||||
const query = `
|
||||
SELECT
|
||||
csv.id as id,
|
||||
@@ -47,6 +66,7 @@ router.get('/transactions/:timeRange', authenticateToken, async (req, res) => {
|
||||
LEFT JOIN fibdash.Kreditor k ON csv.kontonummer_iban = k.iban
|
||||
LEFT JOIN fibdash.BankingAccountTransactions bat ON csv.id = bat.csv_transaction_id
|
||||
LEFT JOIN fibdash.Kreditor ak ON bat.assigned_kreditor_id = ak.id
|
||||
${timeWhereClause}
|
||||
|
||||
UNION ALL
|
||||
|
||||
@@ -84,6 +104,12 @@ router.get('/transactions/:timeRange', authenticateToken, async (req, res) => {
|
||||
WHERE ABS(csv.numeric_amount - jtl.fBetrag) < 0.01
|
||||
AND ABS(DATEDIFF(day, csv.parsed_date, jtl.dBuchungsdatum)) <= 1
|
||||
)
|
||||
${timeRange.includes('-Q') ?
|
||||
`AND YEAR(jtl.dBuchungsdatum) = ${timeRange.split('-Q')[0]} AND MONTH(jtl.dBuchungsdatum) BETWEEN ${(parseInt(timeRange.split('-Q')[1], 10) - 1) * 3 + 1} AND ${(parseInt(timeRange.split('-Q')[1], 10) - 1) * 3 + 3}` :
|
||||
timeRange.length === 4 ?
|
||||
`AND YEAR(jtl.dBuchungsdatum) = ${timeRange}` :
|
||||
`AND YEAR(jtl.dBuchungsdatum) = ${timeRange.split('-')[0]} AND MONTH(jtl.dBuchungsdatum) = ${parseInt(timeRange.split('-')[1], 10)}`
|
||||
}
|
||||
|
||||
ORDER BY parsed_date DESC
|
||||
`;
|
||||
@@ -163,213 +189,32 @@ router.get('/transactions/:timeRange', authenticateToken, async (req, res) => {
|
||||
links: [...new Set(transaction.links.map(l => JSON.stringify(l)))].map(l => JSON.parse(l))
|
||||
}));
|
||||
|
||||
let filteredTransactions = [];
|
||||
|
||||
if (timeRange.includes('-Q')) {
|
||||
const [year, quarterPart] = timeRange.split('-Q');
|
||||
const quarter = parseInt(quarterPart, 10);
|
||||
const startMonth = (quarter - 1) * 3 + 1;
|
||||
const endMonth = startMonth + 2;
|
||||
|
||||
filteredTransactions = transactions.filter(t => {
|
||||
if (!t.monthYear) return false;
|
||||
const [tYear, tMonth] = t.monthYear.split('-');
|
||||
const monthNum = parseInt(tMonth, 10);
|
||||
return tYear === year && monthNum >= startMonth && monthNum <= endMonth;
|
||||
});
|
||||
} else if (timeRange.length === 4) {
|
||||
filteredTransactions = transactions.filter(t => {
|
||||
if (!t.monthYear) return false;
|
||||
const [tYear] = t.monthYear.split('-');
|
||||
return tYear === timeRange;
|
||||
});
|
||||
} else {
|
||||
filteredTransactions = transactions.filter(t => t.monthYear === timeRange);
|
||||
}
|
||||
|
||||
const monthTransactions = filteredTransactions
|
||||
// Transactions are already filtered by the SQL query, so we just need to sort them
|
||||
const monthTransactions = transactions
|
||||
.sort((a, b) => b.parsedDate - a.parsedDate);
|
||||
|
||||
// Get JTL transactions for comparison
|
||||
let jtlTransactions = [];
|
||||
let jtlDatabaseAvailable = false;
|
||||
try {
|
||||
jtlTransactions = await getJTLTransactions();
|
||||
jtlDatabaseAvailable = true;
|
||||
console.log('DEBUG: JTL database connected, found', jtlTransactions.length, 'transactions');
|
||||
} catch (error) {
|
||||
console.log('JTL database not available, continuing without JTL data:', error.message);
|
||||
jtlDatabaseAvailable = false;
|
||||
}
|
||||
|
||||
// Filter JTL transactions for the selected time period
|
||||
let jtlMonthTransactions = [];
|
||||
|
||||
if (timeRange.includes('-Q')) {
|
||||
const [year, quarterPart] = timeRange.split('-Q');
|
||||
const quarter = parseInt(quarterPart, 10);
|
||||
const startMonth = (quarter - 1) * 3 + 1;
|
||||
const endMonth = startMonth + 2;
|
||||
|
||||
jtlMonthTransactions = jtlTransactions.filter(jtl => {
|
||||
const jtlDate = new Date(jtl.dBuchungsdatum);
|
||||
const jtlMonth = jtlDate.getMonth() + 1;
|
||||
return jtlDate.getFullYear() === parseInt(year, 10) &&
|
||||
jtlMonth >= startMonth && jtlMonth <= endMonth;
|
||||
});
|
||||
} else if (timeRange.length === 4) {
|
||||
jtlMonthTransactions = jtlTransactions.filter(jtl => {
|
||||
const jtlDate = new Date(jtl.dBuchungsdatum);
|
||||
return jtlDate.getFullYear() === parseInt(timeRange, 10);
|
||||
});
|
||||
} else {
|
||||
const [year, month] = timeRange.split('-');
|
||||
jtlMonthTransactions = jtlTransactions.filter(jtl => {
|
||||
const jtlDate = new Date(jtl.dBuchungsdatum);
|
||||
return jtlDate.getFullYear() === parseInt(year, 10) &&
|
||||
jtlDate.getMonth() === parseInt(month, 10) - 1;
|
||||
});
|
||||
}
|
||||
|
||||
// Get Kreditor information for IBAN lookup
|
||||
let kreditorData = [];
|
||||
try {
|
||||
const kreditorQuery = `SELECT id, iban, name, kreditorId, is_banking FROM fibdash.Kreditor`;
|
||||
const kreditorResult = await executeQuery(kreditorQuery);
|
||||
kreditorData = kreditorResult.recordset || [];
|
||||
} catch (error) {
|
||||
console.log('Kreditor database not available, continuing without Kreditor data');
|
||||
}
|
||||
|
||||
// Add JTL status and Kreditor information to each CSV transaction
|
||||
const transactionsWithJTL = monthTransactions.map((transaction, index) => {
|
||||
const amount = transaction.numericAmount;
|
||||
const transactionDate = transaction.parsedDate;
|
||||
|
||||
if (index === 0) {
|
||||
console.log('DEBUG First CSV transaction:', {
|
||||
amount: amount,
|
||||
transactionDate: transactionDate,
|
||||
jtlMonthTransactionsCount: jtlMonthTransactions.length
|
||||
});
|
||||
if (jtlMonthTransactions.length > 0) {
|
||||
console.log('DEBUG First JTL transaction:', {
|
||||
amount: parseFloat(jtlMonthTransactions[0].fBetrag),
|
||||
date: new Date(jtlMonthTransactions[0].dBuchungsdatum)
|
||||
});
|
||||
}
|
||||
}
|
||||
|
||||
const jtlMatch = jtlMonthTransactions.find(jtl => {
|
||||
const jtlAmount = parseFloat(jtl.fBetrag) || 0;
|
||||
const jtlDate = new Date(jtl.dBuchungsdatum);
|
||||
|
||||
const amountMatch = Math.abs(amount - jtlAmount) < 0.01;
|
||||
const dateMatch = transactionDate && jtlDate &&
|
||||
transactionDate.getFullYear() === jtlDate.getFullYear() &&
|
||||
transactionDate.getMonth() === jtlDate.getMonth() &&
|
||||
transactionDate.getDate() === jtlDate.getDate();
|
||||
|
||||
if (index === 0 && (amountMatch || dateMatch)) {
|
||||
console.log('DEBUG Potential match for first transaction:', {
|
||||
csvAmount: amount,
|
||||
jtlAmount: jtlAmount,
|
||||
amountMatch: amountMatch,
|
||||
csvDate: transactionDate,
|
||||
jtlDate: jtlDate,
|
||||
dateMatch: dateMatch,
|
||||
bothMatch: amountMatch && dateMatch
|
||||
});
|
||||
}
|
||||
|
||||
return amountMatch && dateMatch;
|
||||
});
|
||||
|
||||
const transactionIban = transaction['Kontonummer/IBAN'];
|
||||
const kreditorMatch = transactionIban ? kreditorData.find(k => k.iban === transactionIban) : null;
|
||||
|
||||
return {
|
||||
...transaction,
|
||||
hasJTL: jtlDatabaseAvailable ? !!jtlMatch : undefined,
|
||||
jtlId: jtlMatch ? jtlMatch.kZahlungsabgleichUmsatz : null,
|
||||
isFromCSV: true,
|
||||
jtlDatabaseAvailable,
|
||||
pdfs: jtlMatch ? jtlMatch.pdfs || [] : [],
|
||||
links: jtlMatch ? jtlMatch.links || [] : [],
|
||||
kreditor: kreditorMatch ? {
|
||||
id: kreditorMatch.id,
|
||||
name: kreditorMatch.name,
|
||||
kreditorId: kreditorMatch.kreditorId,
|
||||
iban: kreditorMatch.iban,
|
||||
is_banking: Boolean(kreditorMatch.is_banking)
|
||||
} : null,
|
||||
hasKreditor: !!kreditorMatch
|
||||
};
|
||||
});
|
||||
|
||||
const unmatchedJTLTransactions = jtlMonthTransactions
|
||||
.filter(jtl => {
|
||||
const jtlAmount = parseFloat(jtl.fBetrag) || 0;
|
||||
const jtlDate = new Date(jtl.dBuchungsdatum);
|
||||
|
||||
const hasCSVMatch = monthTransactions.some(transaction => {
|
||||
const amount = transaction.numericAmount;
|
||||
const transactionDate = transaction.parsedDate;
|
||||
|
||||
const amountMatch = Math.abs(amount - jtlAmount) < 0.01;
|
||||
const dateMatch = transactionDate && jtlDate &&
|
||||
transactionDate.getFullYear() === jtlDate.getFullYear() &&
|
||||
transactionDate.getMonth() === jtlDate.getMonth() &&
|
||||
transactionDate.getDate() === jtlDate.getDate();
|
||||
|
||||
return amountMatch && dateMatch;
|
||||
});
|
||||
|
||||
return !hasCSVMatch;
|
||||
})
|
||||
.map(jtl => ({
|
||||
'Buchungstag': new Date(jtl.dBuchungsdatum).toLocaleDateString('de-DE', {
|
||||
day: '2-digit',
|
||||
month: '2-digit',
|
||||
year: '2-digit'
|
||||
}),
|
||||
'Verwendungszweck': jtl.cVerwendungszweck || '',
|
||||
'Buchungstext': 'JTL Transaction',
|
||||
'Beguenstigter/Zahlungspflichtiger': jtl.cName || '',
|
||||
'Kontonummer/IBAN': '',
|
||||
'Betrag': jtl.fBetrag ? jtl.fBetrag.toString().replace('.', ',') : '0,00',
|
||||
numericAmount: parseFloat(jtl.fBetrag) || 0,
|
||||
parsedDate: new Date(jtl.dBuchungsdatum),
|
||||
monthYear: timeRange,
|
||||
hasJTL: true,
|
||||
jtlId: jtl.kZahlungsabgleichUmsatz,
|
||||
isFromCSV: false,
|
||||
isJTLOnly: true,
|
||||
pdfs: jtl.pdfs || [],
|
||||
links: jtl.links || [],
|
||||
kreditor: null,
|
||||
hasKreditor: false
|
||||
}));
|
||||
// Since transactions are already filtered and joined with JTL data in SQL,
|
||||
// we don't need the complex post-processing logic anymore
|
||||
|
||||
const summary = {
|
||||
totalTransactions: filteredTransactions.length,
|
||||
totalIncome: filteredTransactions
|
||||
totalTransactions: transactions.length,
|
||||
totalIncome: transactions
|
||||
.filter(t => t.numericAmount > 0)
|
||||
.reduce((sum, t) => sum + t.numericAmount, 0),
|
||||
totalExpenses: filteredTransactions
|
||||
totalExpenses: transactions
|
||||
.filter(t => t.numericAmount < 0)
|
||||
.reduce((sum, t) => sum + Math.abs(t.numericAmount), 0),
|
||||
netAmount: filteredTransactions.reduce((sum, t) => sum + t.numericAmount, 0),
|
||||
netAmount: transactions.reduce((sum, t) => sum + t.numericAmount, 0),
|
||||
timeRange: timeRange,
|
||||
jtlDatabaseAvailable: true,
|
||||
jtlMatches: filteredTransactions.filter(t => t.hasJTL === true && t.isFromCSV).length,
|
||||
jtlMissing: filteredTransactions.filter(t => t.hasJTL === false && t.isFromCSV).length,
|
||||
jtlOnly: filteredTransactions.filter(t => t.isJTLOnly === true).length,
|
||||
csvOnly: filteredTransactions.filter(t => t.hasJTL === false && t.isFromCSV).length
|
||||
jtlMatches: transactions.filter(t => t.hasJTL === true && t.isFromCSV).length,
|
||||
jtlMissing: transactions.filter(t => t.hasJTL === false && t.isFromCSV).length,
|
||||
jtlOnly: transactions.filter(t => t.isJTLOnly === true).length,
|
||||
csvOnly: transactions.filter(t => t.hasJTL === false && t.isFromCSV).length
|
||||
};
|
||||
|
||||
res.json({
|
||||
transactions: filteredTransactions,
|
||||
transactions: transactions,
|
||||
summary
|
||||
});
|
||||
} catch (error) {
|
||||
|
||||
@@ -44,7 +44,7 @@ module.exports = {
|
||||
new HtmlWebpackPlugin({
|
||||
template: './client/public/index.html',
|
||||
templateParameters: {
|
||||
REACT_APP_GOOGLE_CLIENT_ID: process.env.GOOGLE_CLIENT_ID || 'your_google_client_id_here',
|
||||
REACT_APP_GOOGLE_CLIENT_ID: process.env.GOOGLE_CLIENT_ID,
|
||||
},
|
||||
}),
|
||||
new webpack.DefinePlugin({
|
||||
@@ -75,7 +75,7 @@ module.exports = {
|
||||
},
|
||||
proxy: {
|
||||
'/api': {
|
||||
target: 'http://localhost:5000',
|
||||
target: 'http://localhost:5500',
|
||||
changeOrigin: true,
|
||||
},
|
||||
},
|
||||
|
||||
@@ -85,7 +85,7 @@ module.exports = {
|
||||
},
|
||||
proxy: {
|
||||
'/api': {
|
||||
target: 'http://localhost:5000',
|
||||
target: 'http://localhost:5500',
|
||||
changeOrigin: true,
|
||||
},
|
||||
},
|
||||
@@ -94,4 +94,4 @@ module.exports = {
|
||||
maxAssetSize: 512000,
|
||||
maxEntrypointSize: 512000,
|
||||
},
|
||||
};
|
||||
};
|
||||
|
||||
Reference in New Issue
Block a user