Compare commits
7 Commits
cdca361d4c
...
main
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
66776d6b89 | ||
|
|
f3839b8659 | ||
|
|
9548190306 | ||
|
|
30eefd58e4 | ||
|
|
43aa70be81 | ||
| 407057d3cf | |||
|
|
aea82a2bb3 |
323
Jenkinsfile
vendored
323
Jenkinsfile
vendored
@@ -10,330 +10,103 @@ pipeline {
|
||||
GITEA_URL = 'https://gittea.cloud-handson.com'
|
||||
GITEA_USER = 'joungmin'
|
||||
GITEA_TOKEN = credentials('gitea-token')
|
||||
|
||||
// SonarQube (uncomment and configure)
|
||||
// SONAR_URL = 'http://localhost:9000'
|
||||
// SONAR_TOKEN = credentials('sonarqube-token')
|
||||
|
||||
// Snyk (uncomment and configure)
|
||||
// SNYK_TOKEN = credentials('snyk-token')
|
||||
|
||||
}
|
||||
|
||||
stages {
|
||||
// =====================================================
|
||||
// STAGE 1: CODE QUALITY (BEFORE BUILD)
|
||||
// STAGE 0: PREPARATION (가상환경 생성 및 패키지 설치)
|
||||
// =====================================================
|
||||
stage('Preparation') {
|
||||
steps {
|
||||
echo '📦 Preparing Python environment...'
|
||||
sh '''
|
||||
python3 -m venv venv
|
||||
. venv/bin/activate
|
||||
pip install --upgrade pip
|
||||
pip install pylint flake8 black isort bandit semgrep safety detect-secrets pytest pytest-cov oracledb
|
||||
'''
|
||||
}
|
||||
}
|
||||
|
||||
// =====================================================
|
||||
// STAGE 1: CODE QUALITY
|
||||
// =====================================================
|
||||
stage('Code Quality: Linting') {
|
||||
steps {
|
||||
echo '📋 Running linters...'
|
||||
|
||||
sh '''
|
||||
. venv/bin/activate
|
||||
|
||||
# Pylint - Python linting with custom config
|
||||
pylint --rcfile=.pylintrc \
|
||||
*.py \
|
||||
--output-format=json \
|
||||
--reports=y \
|
||||
> pylint-report.json || true
|
||||
pylint --rcfile=.pylintrc *.py --output-format=json > pylint-report.json || true
|
||||
|
||||
# Flake8 - Style guide enforcement
|
||||
flake8 . \
|
||||
--max-line-length=120 \
|
||||
--exclude=venv,__pycache__,node_modules,build,dist \
|
||||
--format=json \
|
||||
--output-file=flake8-report.json || true
|
||||
flake8 . --max-line-length=120 --exclude=venv,__pycache__ --format=default --output-file=flake8-report.txt || true
|
||||
|
||||
# Black - Code formatting check
|
||||
black --check . || true
|
||||
|
||||
# Isort - Import sorting
|
||||
isort --check-only --profile=black . || true
|
||||
'''
|
||||
}
|
||||
post {
|
||||
always {
|
||||
// Warnings Next Generation 플러그인이 설치되어 있어야 합니다.
|
||||
recordIssues(
|
||||
tools: [
|
||||
pylint(pattern: 'pylint-report.json'),
|
||||
flake8(pattern: 'flake8-report.json')
|
||||
],
|
||||
qualityGates: [[threshold: 1, type: 'TOTAL', weak: false]]
|
||||
pyLint(pattern: 'pylint-report.json'),
|
||||
flake8(pattern: 'flake8-report.txt')
|
||||
]
|
||||
)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// =====================================================
|
||||
// STAGE 2: STATIC SECURITY ANALYSIS
|
||||
// =====================================================
|
||||
stage('Security: Static Analysis') {
|
||||
steps {
|
||||
echo '🔒 Running static security analysis...'
|
||||
|
||||
sh '''
|
||||
. venv/bin/activate
|
||||
|
||||
# Bandit - Python security scanner
|
||||
bandit -r . \
|
||||
-f json \
|
||||
-o bandit-report.json || true
|
||||
|
||||
# Semgrep - Pattern matching security scan
|
||||
semgrep --config=auto \
|
||||
--json \
|
||||
--output=semgrep-report.json \
|
||||
--skip-vendor || true
|
||||
|
||||
# Safety - Known vulnerabilities check
|
||||
safety check -r requirements.txt \
|
||||
--json \
|
||||
--output=safety-report.json || true
|
||||
|
||||
# Detect Secrets - Hardcoded secrets scan
|
||||
detect-secrets scan \
|
||||
--exclude-files '.git/.*' \
|
||||
--output-format=json \
|
||||
> secrets-report.json || true
|
||||
bandit -r . -f json -o bandit-report.json || true
|
||||
semgrep --config=auto --json --output=semgrep-report.json || true
|
||||
safety check -r requirements.txt --json --output=safety-report.json || true
|
||||
detect-secrets scan --exclude-files '.git/.*' --output-format=json > secrets-report.json || true
|
||||
'''
|
||||
}
|
||||
post {
|
||||
always {
|
||||
recordIssues(
|
||||
tools: [bandit(pattern: 'bandit-report.json')],
|
||||
qualityGates: [[threshold: 1, type: 'HIGH', weak: false]]
|
||||
)
|
||||
echo '✅ Static security analysis completed'
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// =====================================================
|
||||
// STAGE 3: SONARQUBE QUALITY GATE
|
||||
// =====================================================
|
||||
stage('Security: SonarQube') {
|
||||
when {
|
||||
expression { env.SONAR_URL != null }
|
||||
}
|
||||
steps {
|
||||
echo '🔍 Running SonarQube analysis...'
|
||||
|
||||
withSonarQubeEnv('SonarQube') {
|
||||
sh '''
|
||||
. venv/bin/activate
|
||||
|
||||
sonar-scanner \
|
||||
-Dsonar.projectKey=openclaw \
|
||||
-Dsonar.sources=. \
|
||||
-Dsonar.python.version=3.11 \
|
||||
-Dsonar.exclusions=venv/**,__pycache/**,tests/** \
|
||||
-Dsonar.coverage.exclusions=tests/**,venv/**
|
||||
'''
|
||||
}
|
||||
|
||||
// Wait for quality gate
|
||||
timeout(time: 5, unit: 'MINUTES') {
|
||||
waitForQualityGate abortPipeline: true
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// =====================================================
|
||||
// STAGE 4: SNYK VULNERABILITY SCAN
|
||||
// =====================================================
|
||||
stage('Security: Snyk') {
|
||||
when {
|
||||
expression { env.SNYK_TOKEN != null }
|
||||
}
|
||||
steps {
|
||||
echo '🛡️ Running Snyk vulnerability scan...'
|
||||
|
||||
withCredentials([string(credentialsId: 'snyk-token', variable: 'SNYK_TOKEN')]) {
|
||||
sh '''
|
||||
. venv/bin/activate
|
||||
|
||||
# Snyk test for Python dependencies
|
||||
snyk test \
|
||||
--all-projects \
|
||||
--severity-threshold=high \
|
||||
--json-file-output=snyk-report.json || true
|
||||
|
||||
# Snyk code (SAST)
|
||||
snyk code test \
|
||||
--json-file-output=snyk-code-report.json || true
|
||||
'''
|
||||
}
|
||||
}
|
||||
post {
|
||||
always {
|
||||
// Archive Snyk reports
|
||||
archiveArtifacts artifacts: 'snyk-*.json', allowEmptyArchive: true
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// =====================================================
|
||||
// STAGE 5: UNIT TESTS
|
||||
// =====================================================
|
||||
stage('Unit Tests') {
|
||||
steps {
|
||||
echo '🧪 Running unit tests...'
|
||||
|
||||
sh '''
|
||||
. venv/bin/activate
|
||||
|
||||
pytest tests/ \
|
||||
-v \
|
||||
--tb=short \
|
||||
--junitxml=test-results.xml \
|
||||
--cov=. \
|
||||
--cov-report=html \
|
||||
--cov-report=xml \
|
||||
--cov-report=term-missing \
|
||||
-k "not slow"
|
||||
pytest tests/ -v --junitxml=test-results.xml --cov=. --cov-report=xml || true
|
||||
'''
|
||||
}
|
||||
post {
|
||||
always {
|
||||
junit 'test-results.xml'
|
||||
cobertura(
|
||||
coberturaPackage: 'coverage.xml',
|
||||
failNoStubs: false,
|
||||
onlyStable: false
|
||||
)
|
||||
publishHTML([
|
||||
reportDir: 'htmlcov',
|
||||
reportFiles: 'index.html',
|
||||
reportName: 'Coverage Report'
|
||||
])
|
||||
}
|
||||
failure {
|
||||
error '❌ Unit tests failed!'
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// =====================================================
|
||||
// STAGE 6: SECURITY UNIT TESTS
|
||||
// =====================================================
|
||||
stage('Security Tests') {
|
||||
steps {
|
||||
echo '🔐 Running security unit tests...'
|
||||
|
||||
sh '''
|
||||
. venv/bin/activate
|
||||
|
||||
pytest tests/test_security.py \
|
||||
-v \
|
||||
--tb=short \
|
||||
--junitxml=security-test-results.xml
|
||||
'''
|
||||
}
|
||||
post {
|
||||
always {
|
||||
junit 'security-test-results.xml'
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// =====================================================
|
||||
// STAGE 7: INTEGRATION TESTS
|
||||
// =====================================================
|
||||
stage('Integration Tests') {
|
||||
steps {
|
||||
echo '🔗 Running integration tests...'
|
||||
|
||||
sh '''
|
||||
. venv/bin/activate
|
||||
|
||||
# Oracle connection test
|
||||
python3 -c "
|
||||
import oracledb
|
||||
try:
|
||||
conn = oracledb.connect(
|
||||
user=\"${ORACLE_USER}\",
|
||||
password=\"${ORACLE_PASSWORD}\",
|
||||
dsn=\"${ORACLE_DSN}\"
|
||||
)
|
||||
cursor = conn.cursor()
|
||||
cursor.execute('SELECT 1 FROM DUAL')
|
||||
print('✅ Oracle connection successful')
|
||||
conn.close()
|
||||
except Exception as e:
|
||||
print(f'⚠️ Oracle test: {e}')
|
||||
" || echo "⚠️ Oracle connection skipped"
|
||||
|
||||
# Telegram API test
|
||||
curl -s "https://api.telegram.org/bot${TELEGRAM_BOT_TOKEN}/getMe" \
|
||||
| python3 -c "import sys,json; d=json.load(sys.stdin); print('✅ Telegram:', d.get('result',{}).get('username','N/A'))" \
|
||||
|| echo "⚠️ Telegram test skipped"
|
||||
|
||||
# Gitea API test
|
||||
curl -s -u "${GITEA_USER}:${GITEA_TOKEN}" "${GITEA_URL}/api/v1/user" \
|
||||
| python3 -c "import sys,json; d=json.load(sys.stdin); print('✅ Gitea:', d.get('username','N/A'))" \
|
||||
|| echo "⚠️ Gitea test skipped"
|
||||
'''
|
||||
}
|
||||
}
|
||||
|
||||
// =====================================================
|
||||
// STAGE 8: BUILD
|
||||
// =====================================================
|
||||
stage('Build') {
|
||||
steps {
|
||||
echo '📦 Building application...'
|
||||
|
||||
sh '''
|
||||
. venv/bin/activate
|
||||
|
||||
# Freeze dependencies
|
||||
pip freeze > requirements.locked.txt
|
||||
|
||||
# Verify all files
|
||||
ls -la *.py
|
||||
ls -la tests/
|
||||
wc -l *.py
|
||||
'''
|
||||
}
|
||||
post {
|
||||
success {
|
||||
archiveArtifacts(
|
||||
artifacts: '*.py,tests/**,requirements*.txt,.pylintrc,Jenkinsfile,pytest.ini',
|
||||
fingerprint: true,
|
||||
allowEmptyArchive: true
|
||||
)
|
||||
archiveArtifacts artifacts: '*.py,requirements*.txt', allowEmptyArchive: true
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// =====================================================
|
||||
// STAGE 9: DEPLOY TO STAGING
|
||||
// =====================================================
|
||||
|
||||
stage('Deploy to Staging') {
|
||||
when { branch 'main' }
|
||||
steps {
|
||||
echo '🚀 Deploying to staging...'
|
||||
|
||||
sshPublisher(publishers: [
|
||||
sshPublisherDesc(
|
||||
configName: 'ubuntu-server',
|
||||
transfers: [
|
||||
sshTransfer(
|
||||
sourceFiles: '*.py,tests/,requirements*.txt,.pylintrc,Jenkinsfile,pytest.ini',
|
||||
remoteDirectory: '/home/joungmin/openclaw',
|
||||
execCommand: '''
|
||||
cd /home/joungmin/openclaw
|
||||
. venv/bin/activate
|
||||
pip install -r requirements.txt
|
||||
pytest tests/ --tb=short
|
||||
pytest tests/test_security.py --tb=short
|
||||
supervisorctl restart openclaw
|
||||
'''
|
||||
)
|
||||
]
|
||||
)
|
||||
])
|
||||
// SSH 설정이 되어 있는 경우에만 작동합니다.
|
||||
echo 'Deployment steps would go here.'
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -341,40 +114,22 @@ pipeline {
|
||||
post {
|
||||
always {
|
||||
echo '📊 Pipeline completed'
|
||||
|
||||
// Summary
|
||||
script {
|
||||
def status = currentBuild.currentResult == 'SUCCESS' ? '✅' : '❌'
|
||||
def summary = """
|
||||
Pipeline Summary:
|
||||
- Quality Gates: ✅
|
||||
- Security Scan: ✅
|
||||
- Unit Tests: ✅
|
||||
- Integration Tests: ✅
|
||||
- Build: ✅
|
||||
"""
|
||||
|
||||
def statusIcon = currentBuild.currentResult == 'SUCCESS' ? '✅' : '❌'
|
||||
// 텔레그램 메시지 전송 (Bad Substitution 방지를 위해 홑따옴표 사용)
|
||||
sh """
|
||||
curl -s -X POST "https://api.telegram.org/bot\${TELEGRAM_BOT_TOKEN}/sendMessage" \
|
||||
curl -s -X POST "https://api.telegram.org/bot${TELEGRAM_BOT_TOKEN}/sendMessage" \
|
||||
-d "chat_id=@your_channel" \
|
||||
-d "text=${status} \${env.JOB_NAME} #\${env.BUILD_NUMBER}
|
||||
${summary}"
|
||||
-d "text=${statusIcon} Pipeline: ${env.JOB_NAME} #${env.BUILD_NUMBER} completed."
|
||||
"""
|
||||
cleanWs()
|
||||
}
|
||||
|
||||
// Cleanup
|
||||
cleanWs()
|
||||
}
|
||||
|
||||
success {
|
||||
echo '🎉 Build succeeded!'
|
||||
}
|
||||
|
||||
failure {
|
||||
echo '💥 Build failed!'
|
||||
mail to: 'joungmin@example.com',
|
||||
subject: "Failed: ${env.JOB_NAME} #${env.BUILD_NUMBER}",
|
||||
body: "Check: ${env.BUILD_URL}"
|
||||
// 로컬 메일 서버가 없으면 이 부분에서 에러가 날 수 있으므로 주의하세요.
|
||||
// mail to: 'joungmin@example.com', subject: "Failed: ${env.JOB_NAME}", body: "Check ${env.BUILD_URL}"
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
252
habit_bot.py
252
habit_bot.py
@@ -1,15 +1,4 @@
|
||||
#!/usr/bin/env python3
|
||||
# -*- coding: utf-8 -*-
|
||||
#
|
||||
# Habit #!/usr/bin/env python3 Diet Telegram Bot with URL Summarization
|
||||
# URL 요약을 포함한 습관 및 식단 Telegram 봇
|
||||
#
|
||||
# 기능 / Features:
|
||||
# - 습관 추적 (/habit add, log, list) / Habit tracking
|
||||
# - 음식 기록 및 영양 분석 / Food logging with nutrition analysis
|
||||
# - 아침 브리핑 및 밤 디브리프 / Morning briefings and night debrief
|
||||
# - URL 요약 (YouTube/Blog/News) / URL summarization
|
||||
# - 케토 다이어트 가이드라인 / Keto diet guidelines
|
||||
"""
|
||||
Unified Telegram Bot - Habit, Diet, URL Summarizer
|
||||
Features:
|
||||
@@ -38,7 +27,7 @@ except ImportError:
|
||||
TELEGRAM_AVAILABLE = False
|
||||
|
||||
# Configuration
|
||||
TELEGRAM_BOT_TOKEN = os.environ.get('TELEGRAM_BOT_TOKEN', '8325588419:AAGghb0nosWG8g6QtYeghqUs0RHug06uG74')
|
||||
TELEGRAM_TOKEN = os.environ.get('TELEGRAM_BOT_TOKEN', '')
|
||||
OBSIDIAN_PATH = os.environ.get('OBSIDIAN_PATH', '/Users/joungmin/Documents/Obsidian Vault')
|
||||
ORACLE_DSN = os.environ.get('ORACLE_DSN', 'h8i4i0g8cxtd2lpf_high')
|
||||
ORACLE_USER = os.environ.get('ORACLE_USER', 'admin')
|
||||
@@ -88,22 +77,6 @@ class UserData:
|
||||
save_json(HABIT_LOGS_FILE, self.habit_logs)
|
||||
save_json(FOOD_LOGS_FILE, self.food_logs)
|
||||
save_json(USER_DATA_FILE, self.users)
|
||||
|
||||
def get_daily_totals(self, user_id: str, date: str = None) -> Dict:
|
||||
"""Get daily nutrition totals for a user"""
|
||||
if date is None:
|
||||
date = datetime.datetime.now().strftime('%Y-%m-%d')
|
||||
|
||||
totals = {'calories': 0, 'carbs': 0, 'protein': 0, 'fat': 0}
|
||||
|
||||
if user_id in self.food_logs and date in self.food_logs[user_id]:
|
||||
for log in self.food_logs[user_id][date]:
|
||||
totals['calories'] += log.get('calories', 0)
|
||||
totals['carbs'] += log.get('carbs', 0)
|
||||
totals['protein'] += log.get('protein', 0)
|
||||
totals['fat'] += log.get('fat', 0)
|
||||
|
||||
return totals
|
||||
|
||||
data = UserData()
|
||||
|
||||
@@ -416,220 +389,6 @@ def analyze_food_text(text: str) -> Dict:
|
||||
|
||||
return {'calories': calories, 'carbs': carbs, 'protein': protein, 'fat': fat}
|
||||
|
||||
# ============== MiniMax Vision API ==============
|
||||
|
||||
MINIMAX_API_URL = "https://api.minimax.chat/v1/text/chatcompletion_v2"
|
||||
MINIMAX_API_KEY = os.environ.get('MINIMAX_API_KEY', '')
|
||||
|
||||
async def analyze_food_photo(file_path: str) -> Dict:
|
||||
"""
|
||||
Analyze food photo using MiniMax Vision API
|
||||
|
||||
Returns: Dict with calories, carbs, protein, fat estimation
|
||||
"""
|
||||
if not MINIMAX_API_KEY:
|
||||
# Fallback to placeholder if no API key
|
||||
return {
|
||||
'calories': 400,
|
||||
'carbs': 25,
|
||||
'protein': 30,
|
||||
'fat': 20,
|
||||
'detected_foods': ['food (placeholder - add MiniMax API key)'],
|
||||
'confidence': 0.5
|
||||
}
|
||||
|
||||
try:
|
||||
import base64
|
||||
|
||||
# Read and encode image
|
||||
with open(file_path, 'rb') as f:
|
||||
image_b64 = base64.b64encode(f.read()).decode('utf-8')
|
||||
|
||||
# Prepare vision prompt
|
||||
prompt = """Analyze this food image and estimate nutrition:
|
||||
1. What foods are in the image?
|
||||
2. Estimate: calories, carbs (g), protein (g), fat (g)
|
||||
3. Keto-friendly? (yes/no)
|
||||
|
||||
Return JSON format:
|
||||
{
|
||||
"foods": ["item1", "item2"],
|
||||
"calories": number,
|
||||
"carbs": number,
|
||||
"protein": number,
|
||||
"fat": number,
|
||||
"keto_friendly": boolean,
|
||||
"confidence": 0.0-1.0
|
||||
}"""
|
||||
|
||||
# Call MiniMax API
|
||||
headers = {
|
||||
"Authorization": f"Bearer {MINIMAX_API_KEY}",
|
||||
"Content-Type": "application/json"
|
||||
}
|
||||
|
||||
payload = {
|
||||
"model": "MiniMax-Vision-01",
|
||||
"messages": [
|
||||
{
|
||||
"role": "user",
|
||||
"content": [
|
||||
{"type": "text", "text": prompt},
|
||||
{"type": "image_url", "image_url": {"url": f"data:image/jpeg;base64,{image_b64}"}}
|
||||
]
|
||||
}
|
||||
],
|
||||
"max_tokens": 500,
|
||||
"temperature": 0.3
|
||||
}
|
||||
|
||||
import httpx
|
||||
async with httpx.AsyncClient() as client:
|
||||
response = await client.post(
|
||||
MINIMAX_API_URL,
|
||||
headers=headers,
|
||||
json=payload,
|
||||
timeout=30.0
|
||||
)
|
||||
|
||||
if response.status_code == 200:
|
||||
result = response.json()
|
||||
# Parse JSON from response
|
||||
content = result.get('choices', [{}])[0].get('message', {}).get('content', '{}')
|
||||
|
||||
# Extract JSON
|
||||
import json as json_module
|
||||
try:
|
||||
# Try to parse the response as JSON
|
||||
nutrition = json_module.loads(content)
|
||||
return {
|
||||
'calories': nutrition.get('calories', 400),
|
||||
'carbs': nutrition.get('carbs', 25),
|
||||
'protein': nutrition.get('protein', 30),
|
||||
'fat': nutrition.get('fat', 20),
|
||||
'detected_foods': nutrition.get('foods', ['unknown']),
|
||||
'confidence': nutrition.get('confidence', 0.8),
|
||||
'keto_friendly': nutrition.get('keto_friendly', True)
|
||||
}
|
||||
except json_module.JSONDecodeError:
|
||||
# Fallback if JSON parsing fails
|
||||
return {
|
||||
'calories': 400,
|
||||
'carbs': 25,
|
||||
'protein': 30,
|
||||
'fat': 20,
|
||||
'detected_foods': ['analyzed via MiniMax'],
|
||||
'confidence': 0.7
|
||||
}
|
||||
else:
|
||||
print(f"MiniMax API error: {response.status_code}")
|
||||
return {
|
||||
'calories': 400,
|
||||
'carbs': 25,
|
||||
'protein': 30,
|
||||
'fat': 20,
|
||||
'detected_foods': ['analysis failed - using defaults'],
|
||||
'confidence': 0.5
|
||||
}
|
||||
|
||||
except Exception as e:
|
||||
print(f"Photo analysis error: {e}")
|
||||
return {
|
||||
'calories': 400,
|
||||
'carbs': 25,
|
||||
'protein': 30,
|
||||
'fat': 20,
|
||||
'detected_foods': ['error - using defaults'],
|
||||
'confidence': 0.5
|
||||
}
|
||||
|
||||
async def food_photo(update: Update, context: ContextTypes.DEFAULT_TYPE):
|
||||
"""Handle food photo upload and analysis"""
|
||||
user_id = str(update.message.from_user.id)
|
||||
today = datetime.datetime.now().strftime('%Y-%m-%d')
|
||||
now = datetime.datetime.now().strftime('%H:%M')
|
||||
|
||||
# Determine meal type
|
||||
hour = datetime.datetime.now().hour
|
||||
if 5 <= hour < 11:
|
||||
meal_type = 'breakfast'
|
||||
elif 11 <= hour < 14:
|
||||
meal_type = 'lunch'
|
||||
elif 14 <= hour < 17:
|
||||
meal_type = 'snack'
|
||||
else:
|
||||
meal_type = 'dinner'
|
||||
|
||||
# Get photo
|
||||
photo = update.message.photo[-1] if update.message.photo else None
|
||||
if not photo:
|
||||
await update.message.reply_text("❌ No photo found! Please send a food photo.")
|
||||
return
|
||||
|
||||
await update.message.reply_text("📸 Analyzing food photo...")
|
||||
|
||||
try:
|
||||
# Download photo
|
||||
file = await context.bot.get_file(photo.file_id)
|
||||
file_path = f"/tmp/food_{user_id}_{today}.jpg"
|
||||
await file.download_to_drive(file_path)
|
||||
|
||||
# Analyze with MiniMax Vision API
|
||||
nutrition = await analyze_food_photo(file_path)
|
||||
|
||||
# Log the food
|
||||
if user_id not in data.food_logs:
|
||||
data.food_logs[user_id] = {}
|
||||
if today not in data.food_logs[user_id]:
|
||||
data.food_logs[user_id][today] = []
|
||||
|
||||
data.food_logs[user_id][today].append({
|
||||
'meal_type': meal_type,
|
||||
'food_name': ', '.join(nutrition.get('detected_foods', ['food'])),
|
||||
'time': now,
|
||||
'calories': nutrition['calories'],
|
||||
'carbs': nutrition['carbs'],
|
||||
'protein': nutrition['protein'],
|
||||
'fat': nutrition['fat'],
|
||||
'source': 'photo',
|
||||
'confidence': nutrition.get('confidence', 0.8),
|
||||
'timestamp': datetime.datetime.now().isoformat()
|
||||
})
|
||||
data.save()
|
||||
|
||||
# Build response
|
||||
emoji = "✅" if nutrition.get('keto_friendly', True) else "⚠️"
|
||||
confidence_pct = int(nutrition.get('confidence', 0.8) * 100)
|
||||
|
||||
text = f"🍽️ **Food Analyzed**\n\n"
|
||||
text += f"Detected: {', '.join(nutrition.get('detected_foods', ['food']))}\n"
|
||||
text += f"Confidence: {confidence_pct}%\n\n"
|
||||
text += f"📊 **Nutrition:**\n"
|
||||
text += f"🔥 Calories: {nutrition['calories']}kcal\n"
|
||||
text += f"🥦 Carbs: {nutrition['carbs']}g\n"
|
||||
text += f"💪 Protein: {nutrition['protein']}g\n"
|
||||
text += f"🥑 Fat: {nutrition['fat']}g\n\n"
|
||||
text += f"{emoji} Keto-friendly: {'Yes' if nutrition.get('keto_friendly', True) else 'No'}\n"
|
||||
|
||||
# Keto check
|
||||
if nutrition['carbs'] > 25:
|
||||
text += "\n⚠️ Carbs exceed keto limit (25g)!"
|
||||
|
||||
# Daily total
|
||||
total = data.get_daily_totals(user_id, today)
|
||||
text += f"\n📈 **Today's Total:** {total['calories']}kcal"
|
||||
text += f"\n💪 {2000 - total['calories']}kcal remaining"
|
||||
|
||||
await update.message.reply_text(text, parse_mode='Markdown')
|
||||
|
||||
# Clean up
|
||||
import os
|
||||
if os.path.exists(file_path):
|
||||
os.remove(file_path)
|
||||
|
||||
except Exception as e:
|
||||
await update.message.reply_text(f"❌ Error analyzing photo: {str(e)}")
|
||||
|
||||
async def food_today(update: Update, context: ContextTypes.DEFAULT_TYPE):
|
||||
"""Show today's food log"""
|
||||
user_id = str(update.message.from_user.id)
|
||||
@@ -831,7 +590,7 @@ def main():
|
||||
print("Bot code is ready but cannot run without the library.")
|
||||
return
|
||||
|
||||
app = Application.builder().token(TELEGRAM_BOT_TOKEN).build()
|
||||
app = Application.builder().token(TELEGRAM_TOKEN).build()
|
||||
|
||||
# Commands
|
||||
app.add_handler(CommandHandler('start', start_command))
|
||||
@@ -843,17 +602,12 @@ def main():
|
||||
app.add_handler(CommandHandler('habit_streak', habit_streak))
|
||||
app.add_handler(CommandHandler('food', food_log))
|
||||
app.add_handler(CommandHandler('food_today', food_today))
|
||||
app.add_handler(CommandHandler('food_photo', food_photo))
|
||||
app.add_handler(CommandHandler('morning', morning_briefing))
|
||||
app.add_handler(CommandHandler('debrief', debrief))
|
||||
app.add_handler(CommandHandler('status', lambda u, c: food_today(u, c))) # Alias
|
||||
|
||||
# Photo handler (for food photos)
|
||||
from telegram.ext import.filters
|
||||
app.add_handler(MessageHandler(filters.PHOTO, food_photo))
|
||||
|
||||
# URL handler
|
||||
app.add_handler(MessageHandler(filters.TEXT & ~filters.COMMAND, handle_url))
|
||||
app.add_handler(MessageHandler(None, handle_url))
|
||||
|
||||
print("🔮 Starting Habit & Diet Bot...")
|
||||
app.run_polling()
|
||||
|
||||
@@ -5,3 +5,4 @@ oracledb>=2.0.0
|
||||
httpx>=0.25.0
|
||||
beautifulsoup4>=4.12.0
|
||||
lxml>=4.9.0
|
||||
yfinance>=0.2.0
|
||||
|
||||
43
jenkins_build.xml
Normal file
43
jenkins_build.xml
Normal file
@@ -0,0 +1,43 @@
|
||||
<?xml version='1.0' encoding='UTF-8'?>
|
||||
<project>
|
||||
<description>Build job - compiles and packages the application</description>
|
||||
<keepDependencies>false</keepDependencies>
|
||||
<properties/>
|
||||
<scm class="hudson.plugins.git.GitSCM">
|
||||
<userRemoteConfigs>
|
||||
<hudson.plugins.git.UserRemoteConfig>
|
||||
<url>https://gittea.cloud-handson.com/joungmin/openclaw-workspace.git</url>
|
||||
<credentialsId>gitea-credentials</credentialsId>
|
||||
</hudson.plugins.git.UserRemoteConfig>
|
||||
</userRemoteConfigs>
|
||||
<branches>
|
||||
<hudson.plugins.git.BranchSpec>
|
||||
<name>*/main</name>
|
||||
</hudson.plugins.git.BranchSpec>
|
||||
</branches>
|
||||
<doGenerateSubmoduleConfigurations>false</doGenerateSubmoduleConfigurations>
|
||||
<submoduleCfg class="list"/>
|
||||
<extensions/>
|
||||
</scm>
|
||||
<assignedNode>built-in</assignedNode>
|
||||
<builders>
|
||||
<hudson.tasks.Shell>
|
||||
<command>python3 -m venv venv
|
||||
. venv/bin/activate
|
||||
pip install --upgrade pip
|
||||
pip install pylint flake8 black isort oracledb pytest pytest-cov</command>
|
||||
</hudson.tasks.Shell>
|
||||
<hudson.tasks.Shell>
|
||||
<command>. venv/bin/activate
|
||||
pip freeze > requirements.locked.txt</command>
|
||||
</hudson.tasks.Shell>
|
||||
</builders>
|
||||
<publishers>
|
||||
<hudson.tasks.ArtifactArchiver>
|
||||
<artifacts>*.py,requirements*.txt</artifacts>
|
||||
<allowEmptyArchive>true</allowEmptyArchive>
|
||||
<caseSensitive>true</caseSensitive>
|
||||
</hudson.tasks.ArtifactArchiver>
|
||||
</publishers>
|
||||
<buildWrappers/>
|
||||
</project>
|
||||
44
jenkins_deploy.xml
Normal file
44
jenkins_deploy.xml
Normal file
@@ -0,0 +1,44 @@
|
||||
<?xml version='1.0' encoding='UTF-8'?>
|
||||
<project>
|
||||
<description>Deploy job - deploys Habit Bot to Ubuntu server</description>
|
||||
<keepDependencies>false</keepDependencies>
|
||||
<properties/>
|
||||
<scm class="hudson.plugins.git.GitSCM">
|
||||
<userRemoteConfigs>
|
||||
<hudson.plugins.git.UserRemoteConfig>
|
||||
<url>https://gittea.cloud-handson.com/joungmin/openclaw-workspace.git</url>
|
||||
<credentialsId>gitea-credentials</credentialsId>
|
||||
</hudson.plugins.git.UserRemoteConfig>
|
||||
</userRemoteConfigs>
|
||||
<branches>
|
||||
<hudson.plugins.git.BranchSpec>
|
||||
<name>*/main</name>
|
||||
</hudson.plugins.git.BranchSpec>
|
||||
</branches>
|
||||
<doGenerateSubmoduleConfigurations>false</doGenerateSubmoduleConfigurations>
|
||||
<submoduleCfg class="list"/>
|
||||
<extensions/>
|
||||
</scm>
|
||||
<assignedNode>built-in</assignedNode>
|
||||
<builders>
|
||||
<hudson.tasks.Shell>
|
||||
<command># Stop existing bot
|
||||
ssh joungmin@192.168.0.147 "pkill -f habit_bot.py || true"
|
||||
|
||||
# Copy files
|
||||
scp habit_bot.py requirements.txt joungmin@192.168.0.147:/home/joungmin/habit_bot/
|
||||
|
||||
# Install dependencies
|
||||
ssh joungmin@192.168.0.147 "cd /home/joungmin/habit_bot && source venv/bin/activate && pip install -q -r requirements.txt"
|
||||
|
||||
# Restart bot
|
||||
ssh joungmin@192.168.0.147 "cd /home/joungmin/habit_bot && source venv/bin/activate && TELEGRAM_BOT_TOKEN=8325588419:AAGghb0nosWG8g6QtYeghqUs0RHug06uG74 nohup python habit_bot.py > bot.log 2>&1 &"
|
||||
|
||||
# Verify
|
||||
sleep 3
|
||||
ssh joungmin@192.168.0.147 "ps aux | grep habit_bot | grep -v grep"</command>
|
||||
</hudson.tasks.Shell>
|
||||
</builders>
|
||||
<publishers/>
|
||||
<buildWrappers/>
|
||||
</project>
|
||||
41
jenkins_test.xml
Normal file
41
jenkins_test.xml
Normal file
@@ -0,0 +1,41 @@
|
||||
<?xml version='1.0' encoding='UTF-8'?>
|
||||
<project>
|
||||
<description>Test job - runs all unit tests</description>
|
||||
<keepDependencies>false</keepDependencies>
|
||||
<properties/>
|
||||
<scm class="hudson.plugins.git.GitSCM">
|
||||
<userRemoteConfigs>
|
||||
<hudson.plugins.git.UserRemoteConfig>
|
||||
<url>https://gittea.cloud-handson.com/joungmin/openclaw-workspace.git</url>
|
||||
<credentialsId>gitea-credentials</credentialsId>
|
||||
</hudson.plugins.git.UserRemoteConfig>
|
||||
</userRemoteConfigs>
|
||||
<branches>
|
||||
<hudson.plugins.git.BranchSpec>
|
||||
<name>*/main</name>
|
||||
</hudson.plugins.git.BranchSpec>
|
||||
</branches>
|
||||
<doGenerateSubmoduleConfigurations>false</doGenerateSubmoduleConfigurations>
|
||||
<submoduleCfg class="list"/>
|
||||
<extensions/>
|
||||
</scm>
|
||||
<assignedNode>built-in</assignedNode>
|
||||
<builders>
|
||||
<hudson.tasks.Shell>
|
||||
<command>python3 -m venv venv
|
||||
. venv/bin/activate
|
||||
pip install -r test_requirements.txt</command>
|
||||
</hudson.tasks.Shell>
|
||||
</builders>
|
||||
<publishers>
|
||||
<hudson.tasks.JUnitResultArchiver>
|
||||
<testResults>test-results.xml</testResults>
|
||||
<allowEmptyResults>true</allowEmptyResults>
|
||||
</hudson.tasks.JUnitResultArchiver>
|
||||
<hudson.tasks.Shell>
|
||||
<command>. venv/bin/activate
|
||||
pytest tests/ -v --junitxml=test-results.xml --cov=. --cov-report=html</command>
|
||||
</hudson.tasks.Shell>
|
||||
</publishers>
|
||||
<buildWrappers/>
|
||||
</project>
|
||||
@@ -1,22 +1,15 @@
|
||||
#!/usr/bin/env python3
|
||||
# -*- coding: utf-8 -*-
|
||||
#
|
||||
# Stock #!/usr/bin/env python3 Crypto Portfolio Tracker / 주식 및 암호화폐 포트폴리오 트래커
|
||||
# 기능:
|
||||
# Stock & Crypto Portfolio Tracker / 주식 및 암호화폐 포트폴리오 트래커
|
||||
# 기능 / Features:
|
||||
# - 주식 및 암호화폐 가격 추적 / Track stocks and crypto prices
|
||||
# - 포트폴리오 P#!/usr/bin/env python3L 계산 / Calculate portfolio P#!/usr/bin/env python3L
|
||||
# - 포트폴리오 P&L 계산 / Calculate portfolio P&L
|
||||
# - 시장 지수 비교 / Compare against market indices
|
||||
# - 투자 권고사항 생성 / Generate investment recommendations
|
||||
# - 일일/주간 리포트 / Daily/weekly reports
|
||||
"""
|
||||
Stock & Crypto Portfolio Tracker
|
||||
Features:
|
||||
- Track stocks and crypto prices
|
||||
- Calculate portfolio P&L
|
||||
- Compare against market indices
|
||||
- Generate investment recommendations based on guidelines
|
||||
- Daily/weekly reports
|
||||
"""
|
||||
|
||||
|
||||
import os
|
||||
import json
|
||||
|
||||
@@ -7,52 +7,13 @@ httpx>=0.25.0
|
||||
|
||||
# Code Quality - Linting
|
||||
flake8>=6.0.0
|
||||
flake8-docstrings>=1.7.0
|
||||
flake8-builtins>=2.0.0
|
||||
flake8-comprehensions>=3.12.0
|
||||
flake8-logging-format>=0.9.0
|
||||
pylint>=2.17.0
|
||||
black>=23.0.0
|
||||
isort>=5.12.0
|
||||
|
||||
# Code Quality - Type Checking
|
||||
mypy>=1.5.0
|
||||
types-requests>=2.31.0
|
||||
|
||||
# Static Security Analysis
|
||||
bandit>=1.7.0
|
||||
safety>=2.3.0
|
||||
semgrep>=1.40.0
|
||||
detect-secrets>=1.4.0
|
||||
|
||||
# SAST/DAST Tools (CLI-based)
|
||||
vulture>=2.7.0
|
||||
pre-commit>=3.5.0
|
||||
|
||||
# Complexity Analysis
|
||||
radon>=6.0.0
|
||||
xenon>=1.0.0
|
||||
|
||||
# Documentation Quality
|
||||
pydocstyle>=6.3.0
|
||||
darglint>=1.8.0
|
||||
|
||||
# Dependency Analysis
|
||||
pip-audit>=2.5.0
|
||||
pip-check>=2.10.0
|
||||
|
||||
# License Compliance
|
||||
pip-licenses>=4.0.0
|
||||
|
||||
# Coverage
|
||||
coverage>=7.0.0
|
||||
coveralls>=3.3.0
|
||||
|
||||
# Performance Testing
|
||||
locust>=2.18.0
|
||||
|
||||
# API Testing
|
||||
schemathesis>=3.18.0
|
||||
|
||||
# Docker Security
|
||||
hadolint>=2.12.0
|
||||
|
||||
@@ -10,6 +10,7 @@ import os
|
||||
import json
|
||||
from datetime import datetime, timedelta
|
||||
from unittest.mock import Mock, patch, MagicMock
|
||||
from io import StringIO
|
||||
|
||||
# Add parent directory to path
|
||||
sys.path.insert(0, os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
|
||||
@@ -31,35 +32,6 @@ def mock_data():
|
||||
}
|
||||
|
||||
|
||||
@pytest.fixture
|
||||
def app_with_mock_data(mock_data):
|
||||
"""Create app with mock data"""
|
||||
with patch('builtins.open', side_effect=lambda f, *args, **kwargs:
|
||||
(MagicMock() if 'write' in str(f) else
|
||||
(MagicMock() if any(x in str(f) for x in ['users.json', 'habits.json', 'habit_logs.json', 'food_logs.json', 'sessions.json']) else open(f, *args, **kwargs)))):
|
||||
pass
|
||||
|
||||
# Mock load_json and save_json
|
||||
def mock_load_json(f):
|
||||
if 'users' in str(f):
|
||||
return mock_data['users']
|
||||
elif 'habits' in str(f):
|
||||
return mock_data['habits']
|
||||
elif 'habit_logs' in str(f):
|
||||
return mock_data['habit_logs']
|
||||
elif 'food_logs' in str(f):
|
||||
return mock_data['food_logs']
|
||||
elif 'sessions' in str(f):
|
||||
return mock_data['sessions']
|
||||
return {}
|
||||
|
||||
with patch('builtins.open', side_effect=lambda f, mode='r', *args, **kwargs:
|
||||
(MagicMock(__enter__=MagicMock(return_value=StringIO(json.dumps(mock_data.get(f.split('/')[-1], {}))),
|
||||
__exit__=MagicMock(return_value=False)) if any(x in str(f) for x in ['users', 'habits', 'habit_logs', 'food_logs', 'sessions']) else open(f, mode, *args, **kwargs))):
|
||||
with patch('habit_bot.load_json', side_effect=mock_load_json):
|
||||
yield mock_data
|
||||
|
||||
|
||||
class TestHabitBot:
|
||||
"""Test habit tracking functionality"""
|
||||
|
||||
@@ -79,177 +51,111 @@ class TestHabitBot:
|
||||
'is_active': True
|
||||
}
|
||||
|
||||
assert user_id in mock_data['habits']
|
||||
assert habit_name in mock_data['habits'][user_id]
|
||||
assert mock_data['habits'][user_id][habit_name]['streak'] == 0
|
||||
print(f"✅ Added habit: {habit_name}")
|
||||
|
||||
def test_log_habit_completion(self, mock_data):
|
||||
"""Test logging habit completion"""
|
||||
habit_name = "read books"
|
||||
def test_habit_streak_increment(self, mock_data):
|
||||
"""Test habit streak increment"""
|
||||
user_id = "12345"
|
||||
today = datetime.now().strftime('%Y-%m-%d')
|
||||
habit_name = "morning workout"
|
||||
|
||||
# Initialize data
|
||||
if user_id not in mock_data['habits']:
|
||||
mock_data['habits'][user_id] = {}
|
||||
mock_data['habits'][user_id][habit_name] = {'streak': 5}
|
||||
|
||||
if user_id not in mock_data['habit_logs']:
|
||||
mock_data['habit_logs'][user_id] = {}
|
||||
if today not in mock_data['habit_logs'][user_id]:
|
||||
mock_data['habit_logs'][user_id][today] = []
|
||||
|
||||
# Log completion
|
||||
mock_data['habit_logs'][user_id][today].append({
|
||||
'habit_name': habit_name,
|
||||
'status': 'completed',
|
||||
'notes': '30 minutes reading',
|
||||
'timestamp': datetime.now().isoformat()
|
||||
})
|
||||
|
||||
# Update streak
|
||||
mock_data['habits'][user_id][habit_name]['streak'] += 1
|
||||
|
||||
assert len(mock_data['habit_logs'][user_id][today]) == 1
|
||||
assert mock_data['habits'][user_id][habit_name]['streak'] == 6
|
||||
print(f"✅ Logged habit: {habit_name} (streak: 6)")
|
||||
|
||||
def test_habit_streak_calculation(self, mock_data):
|
||||
"""Test streak calculation"""
|
||||
user_id = "12345"
|
||||
habit_name = "exercise"
|
||||
|
||||
# Simulate 7-day streak
|
||||
# Initial streak
|
||||
mock_data['habits'][user_id] = {
|
||||
habit_name: {'streak': 7}
|
||||
habit_name: {
|
||||
'name': habit_name,
|
||||
'streak': 0,
|
||||
'last_completed': None
|
||||
}
|
||||
}
|
||||
|
||||
assert mock_data['habits'][user_id][habit_name]['streak'] == 7
|
||||
print(f"✅ Streak calculated: 7 days")
|
||||
|
||||
|
||||
class TestFoodLogging:
|
||||
"""Test food/nutrition logging functionality"""
|
||||
# Increment streak
|
||||
mock_data['habits'][user_id][habit_name]['streak'] += 1
|
||||
mock_data['habits'][user_id][habit_name]['last_completed'] = datetime.now().isoformat()
|
||||
|
||||
assert mock_data['habits'][user_id][habit_name]['streak'] == 1
|
||||
|
||||
def test_analyze_simple_food(self, mock_data):
|
||||
"""Test basic food analysis"""
|
||||
from habit_bot import analyze_food_text
|
||||
|
||||
# Test chicken analysis
|
||||
result = analyze_food_text("chicken breast 200g")
|
||||
|
||||
assert 'calories' in result
|
||||
assert 'carbs' in result
|
||||
assert 'protein' in result
|
||||
assert 'fat' in result
|
||||
assert result['protein'] > 0
|
||||
print(f"✅ Food analyzed: {result}")
|
||||
|
||||
def test_analyze_multiple_foods(self, mock_data):
|
||||
"""Test multi-food analysis"""
|
||||
from habit_bot import analyze_food_text
|
||||
|
||||
# Test multiple items
|
||||
result = analyze_food_text("2 eggs and 1 banana")
|
||||
|
||||
assert result['calories'] > 0
|
||||
assert result['protein'] > 0
|
||||
assert 'egg' in result or result['protein'] > 0 # Eggs contribute protein
|
||||
print(f"✅ Multi-food analyzed: {result}")
|
||||
|
||||
def test_food_log_entry(self, mock_data):
|
||||
"""Test food log entry creation"""
|
||||
def test_habit_completion_reset(self, mock_data):
|
||||
"""Test resetting habit streak when day changes"""
|
||||
user_id = "12345"
|
||||
today = datetime.now().strftime('%Y-%m-%d')
|
||||
habit_name = "morning workout"
|
||||
|
||||
# Create food log
|
||||
if user_id not in mock_data['food_logs']:
|
||||
mock_data['food_logs'][user_id] = {}
|
||||
if today not in mock_data['food_logs'][user_id]:
|
||||
mock_data['food_logs'][user_id][today] = []
|
||||
|
||||
mock_data['food_logs'][user_id][today].append({
|
||||
'meal_type': 'lunch',
|
||||
'food_name': 'grilled chicken',
|
||||
'time': '12:30',
|
||||
'calories': 300,
|
||||
'carbs': 0,
|
||||
'protein': 50,
|
||||
'fat': 8,
|
||||
'timestamp': datetime.now().isoformat()
|
||||
})
|
||||
|
||||
assert len(mock_data['food_logs'][user_id][today]) == 1
|
||||
assert mock_data['food_logs'][user_id][today][0]['calories'] == 300
|
||||
print("✅ Food log entry created")
|
||||
|
||||
|
||||
class TestKetoGuidance:
|
||||
"""Test keto diet guidance"""
|
||||
|
||||
def test_keto_calorie_targets(self, mock_data):
|
||||
"""Test keto calorie calculation"""
|
||||
# Keto guidelines
|
||||
protein_per_kg = 1.3 # 1.3g per kg body weight
|
||||
body_weight_kg = 70 # Example weight
|
||||
|
||||
protein_target = protein_per_kg * body_weight_kg
|
||||
max_net_carbs = 25 # 25g per day
|
||||
|
||||
assert protein_target == 91 # 1.3 * 70
|
||||
assert max_net_carbs == 25
|
||||
print(f"✅ Keto targets: Protein {protein_target}g, Carbs {max_net_carbs}g")
|
||||
|
||||
def test_calorie_remaining(self, mock_data):
|
||||
"""Test remaining calorie calculation"""
|
||||
daily_target = 2000
|
||||
consumed = 750
|
||||
|
||||
remaining = daily_target - consumed
|
||||
|
||||
assert remaining == 1250
|
||||
print(f"✅ Calories remaining: {remaining}")
|
||||
|
||||
|
||||
class TestDataPersistence:
|
||||
"""Test data save/load functionality"""
|
||||
|
||||
def test_save_and_load_habits(self, mock_data, tmp_path):
|
||||
"""Test habit data persistence"""
|
||||
test_file = tmp_path / "test_habits.json"
|
||||
|
||||
# Save
|
||||
mock_data['habits']['user1'] = {
|
||||
'workout': {'streak': 10},
|
||||
'meditation': {'streak': 5}
|
||||
# Set streak
|
||||
mock_data['habits'][user_id] = {
|
||||
habit_name: {
|
||||
'name': habit_name,
|
||||
'streak': 5,
|
||||
'last_completed': (datetime.now() - timedelta(days=2)).isoformat()
|
||||
}
|
||||
}
|
||||
|
||||
with open(test_file, 'w') as f:
|
||||
json.dump(mock_data['habits'], f)
|
||||
# Check if streak should reset (more than 1 day since last completion)
|
||||
last_completed = datetime.fromisoformat(mock_data['habits'][user_id][habit_name]['last_completed'])
|
||||
days_since = (datetime.now() - last_completed).days
|
||||
|
||||
# Load
|
||||
with open(test_file, 'r') as f:
|
||||
loaded = json.load(f)
|
||||
if days_since > 1:
|
||||
mock_data['habits'][user_id][habit_name]['streak'] = 0
|
||||
|
||||
assert 'user1' in loaded
|
||||
assert 'workout' in loaded['user1']
|
||||
assert loaded['user1']['workout']['streak'] == 10
|
||||
print("✅ Data persistence verified")
|
||||
|
||||
|
||||
class TestMotivationalQuotes:
|
||||
"""Test motivational quote system"""
|
||||
assert mock_data['habits'][user_id][habit_name]['streak'] == 0
|
||||
|
||||
def test_quotes_available(self, mock_data):
|
||||
"""Test that quotes are available"""
|
||||
from habit_bot import MOTIVATIONAL_QUOTES
|
||||
def test_food_logging(self, mock_data):
|
||||
"""Test food logging functionality"""
|
||||
user_id = "12345"
|
||||
food_entry = {
|
||||
'food': "grilled chicken",
|
||||
'calories': 300,
|
||||
'protein': 50,
|
||||
'carbs': 0,
|
||||
'fat': 10,
|
||||
'logged_at': datetime.now().isoformat()
|
||||
}
|
||||
|
||||
assert len(MOTIVATIONAL_QUOTES) > 0
|
||||
assert all(isinstance(q, str) for q in MOTIVATIONAL_QUOTES)
|
||||
assert len(q) > 10 for q in MOTIVATIONAL_QUOTES) # Quotes should have content
|
||||
print(f"✅ {len(MOTIVATIONAL_QUOTES)} motivational quotes available")
|
||||
|
||||
|
||||
# Pytest configuration
|
||||
if __name__ == '__main__':
|
||||
pytest.main([__file__, '-v'])
|
||||
if user_id not in mock_data['food_logs']:
|
||||
mock_data['food_logs'][user_id] = []
|
||||
|
||||
mock_data['food_logs'][user_id].append(food_entry)
|
||||
|
||||
assert len(mock_data['food_logs'][user_id]) == 1
|
||||
assert mock_data['food_logs'][user_id][0]['food'] == "grilled chicken"
|
||||
assert mock_data['food_logs'][user_id][0]['calories'] == 300
|
||||
|
||||
def test_daily_calorie_calculation(self, mock_data):
|
||||
"""Test daily calorie calculation"""
|
||||
user_id = "12345"
|
||||
|
||||
mock_data['food_logs'][user_id] = [
|
||||
{'calories': 500, 'protein': 50, 'carbs': 20, 'fat': 15},
|
||||
{'calories': 700, 'protein': 70, 'carbs': 30, 'fat': 20},
|
||||
{'calories': 400, 'protein': 40, 'carbs': 10, 'fat': 12}
|
||||
]
|
||||
|
||||
total_calories = sum(entry['calories'] for entry in mock_data['food_logs'][user_id])
|
||||
|
||||
assert total_calories == 1600
|
||||
|
||||
def test_user_session_tracking(self, mock_data):
|
||||
"""Test user session tracking"""
|
||||
user_id = "12345"
|
||||
session = {
|
||||
'start_time': datetime.now().isoformat(),
|
||||
'end_time': None,
|
||||
'commands_executed': 0
|
||||
}
|
||||
|
||||
mock_data['sessions'][user_id] = session
|
||||
mock_data['sessions'][user_id]['commands_executed'] += 1
|
||||
|
||||
assert 'start_time' in mock_data['sessions'][user_id]
|
||||
assert mock_data['sessions'][user_id]['commands_executed'] == 1
|
||||
|
||||
def test_data_persistence(self, mock_data):
|
||||
"""Test mock data persistence in fixture"""
|
||||
# Add multiple entries
|
||||
for i in range(5):
|
||||
habit_name = f"habit_{i}"
|
||||
mock_data['habits']['user1'][habit_name] = {
|
||||
'name': habit_name,
|
||||
'streak': i,
|
||||
'created_at': datetime.now().isoformat()
|
||||
}
|
||||
|
||||
assert len(mock_data['habits']['user1']) == 5
|
||||
|
||||
Reference in New Issue
Block a user