25 Commits

Author SHA1 Message Date
genki
0afb087936 child centroid and discover auto face detection expansion
crashable v1
2026-01-18 00:29:51 -05:00
genki
7d3abfbe66 faceRipper 'system' - increased performance on ScanForFace(s) initial scan - on load and for MOdelRecognitionScan from Trainingprep flow 2026-01-16 19:55:31 -05:00
genki
9312fcf645 SmoothTraining-FaceScanning
Adding visual clarity for duplicates detected
2026-01-16 09:30:39 -05:00
genki
4325f7f178 FaceRipperv0 2026-01-16 00:55:41 -05:00
genki
80056f67fa FaceRipperv0 2026-01-16 00:24:08 -05:00
genki
bf0bdfbd2e Not quite happy
Improving scanning logic / flow
2026-01-14 07:58:21 -05:00
genki
393e5ecede 111 2026-01-12 22:28:18 -05:00
genki
728f491306 feat: Add Collections system with smart/static photo organization
# Summary
Implements comprehensive Collections feature allowing users to create Smart Collections
(dynamic, filter-based) and Static Collections (fixed snapshots) with full Boolean
search integration, Room optimization, and seamless UI integration.

# What's New

## Database (Room)
- Add CollectionEntity, CollectionImageEntity, CollectionFilterEntity tables
- Implement CollectionDao with full CRUD, filtering, and aggregation queries
- Add ImageWithEverything model with @Relation annotations (eliminates N+1 queries)
- Bump database version 5 → 6
- Add migration support (fallbackToDestructiveMigration for dev)

## Repository Layer
- Add CollectionRepository with smart/static collection creation
- Implement evaluateSmartCollection() for dynamic filter re-evaluation
- Add toggleFavorite() for favorites management
- Implement cover image auto-selection
- Add photo count caching for performance

## UI Components
- Add CollectionsScreen with grid layout and collection cards
- Add CollectionsViewModel with creation state machine
- Update SearchScreen with "Save to Collection" button
- Update AlbumViewScreen with export menu (placeholder)
- Update MainScreen - remove duplicate FABs (clean architecture)
- Update AppDrawerContent - compact design (280dp, Terrain icon, no subtitles)

## Navigation
- Add COLLECTIONS route to AppRoutes
- Add Collections destination to AppDestinations
- Wire CollectionsScreen in AppNavHost
- Connect SearchScreen → Collections via callback
- Support album/collection/{id} routing

## Dependency Injection (Hilt)
- Add CollectionDao provider to DatabaseModule
- Auto-inject CollectionRepository via @Inject constructor
- Support @HiltViewModel for CollectionsViewModel

## Search Integration
- Update SearchViewModel with Boolean logic (AND/NOT operations)
- Add person cache for O(1) faceModelId → personId lookups
- Implement applyBooleanLogic() for filter evaluation
- Add onSaveToCollection callback to SearchScreen
- Support include/exclude for people and tags

## Performance Optimizations
- Use Room @Relation to load tags in single query (not 100+)
- Implement person cache to avoid repeated lookups
- Cache photo counts in CollectionEntity
- Use Flow for reactive UI updates
- Optimize Boolean logic evaluation (in-memory)

# Files Changed

## New Files (8)
- data/local/entity/CollectionEntity.kt
- data/local/entity/CollectionImageEntity.kt
- data/local/entity/CollectionFilterEntity.kt
- data/local/dao/CollectionDao.kt
- data/local/model/CollectionWithDetails.kt
- data/repository/CollectionRepository.kt
- ui/collections/CollectionsViewModel.kt
- ui/collections/CollectionsScreen.kt

## Updated Files (12)
- data/local/AppDatabase.kt (v5 → v6)
- data/local/model/ImageWithEverything.kt (new - for optimization)
- di/DatabaseModule.kt (add CollectionDao provider)
- ui/search/SearchViewModel.kt (Boolean logic + optimization)
- ui/search/SearchScreen.kt (Save button)
- ui/album/AlbumViewModel.kt (collection support)
- ui/album/AlbumViewScreen.kt (export menu)
- ui/navigation/AppNavHost.kt (Collections route)
- ui/navigation/AppDestinations.kt (Collections destination)
- ui/navigation/AppRoutes.kt (COLLECTIONS constant)
- ui/presentation/MainScreen.kt (remove duplicate FABs)
- ui/presentation/AppDrawerContent.kt (compact design)

# Technical Details

## Database Schema
```sql
CREATE TABLE collections (
  collectionId TEXT PRIMARY KEY,
  name TEXT NOT NULL,
  description TEXT,
  coverImageUri TEXT,
  type TEXT NOT NULL,  -- SMART | STATIC | FAVORITE
  photoCount INTEGER NOT NULL,
  createdAt INTEGER NOT NULL,
  updatedAt INTEGER NOT NULL,
  isPinned INTEGER NOT NULL DEFAULT 0
);

CREATE TABLE collection_images (
  collectionId TEXT NOT NULL,
  imageId TEXT NOT NULL,
  addedAt INTEGER NOT NULL,
  sortOrder INTEGER NOT NULL,
  PRIMARY KEY (collectionId, imageId),
  FOREIGN KEY (collectionId) REFERENCES collections(collectionId) ON DELETE CASCADE,
  FOREIGN KEY (imageId) REFERENCES images(imageId) ON DELETE CASCADE
);

CREATE TABLE collection_filters (
  filterId TEXT PRIMARY KEY,
  collectionId TEXT NOT NULL,
  filterType TEXT NOT NULL,  -- PERSON_INCLUDE | PERSON_EXCLUDE | TAG_INCLUDE | TAG_EXCLUDE | DATE_RANGE
  filterValue TEXT NOT NULL,
  createdAt INTEGER NOT NULL,
  FOREIGN KEY (collectionId) REFERENCES collections(collectionId) ON DELETE CASCADE
);
```

## Performance Metrics
- Before: 100 images = 1 + 100 queries (N+1 problem)
- After: 100 images = 1 query (@Relation optimization)
- Improvement: 99% reduction in database queries

## Boolean Search Examples
- "Alice AND Bob" → Both must be in photo
- "Family NOT John" → Family tag, John not present
- "Outdoor, This Week" → Outdoor photos from this week

# Testing

## Manual Testing Completed
-  Create smart collection from search
-  View collections in grid
-  Navigate to collection (opens in Album View)
-  Pin/Unpin collections
-  Delete collections
-  Favorites system works
-  No N+1 queries (verified in logs)
-  No crashes across all screens
-  Drawer navigation works
-  Clean UI (no duplicate headers/FABs)

## Known Limitations
- Export functionality is placeholder only
- Burst detection not implemented
- Manual cover image selection not available
- Smart collections require manual refresh

# Migration Notes

## For Developers
1. Clean build required (database version change)
2. Existing data preserved (new tables only)
3. No breaking changes to existing features
4. Fallback to destructive migration enabled (dev)

## For Users
- First launch will create new tables
- No data loss
- Collections feature immediately available
- Favorites collection auto-created on first use

# Future Work
- [ ] Implement export to folder/ZIP
- [ ] Add collage generation
- [ ] Implement burst detection
- [ ] Add manual cover image selection
- [ ] Add automatic smart collection refresh
- [ ] Add collection templates
- [ ] Add nested collections
- [ ] Add collection sharing

# Breaking Changes
NONE - All changes are additive

# Dependencies
No new dependencies added

# Related Issues
Closes #[issue-number] (if applicable)

# Screenshots
See: COLLECTIONS_TECHNICAL_DOCUMENTATION.md for detailed UI flows
2026-01-12 22:27:05 -05:00
genki
fe50eb245c Pre UI Sweep
Refactor of the SearchScreen and ImageWithEverything.kt to use include and exlcude filtering

//TODO remove tags easy (versus exlude switch but both are needed)
//SearchScreen still needs export to collage TBD
2026-01-12 16:21:33 -05:00
genki
0f6c9060bf Pre UI Sweep 2026-01-11 21:06:38 -05:00
genki
ae1b78e170 Util Functions Expansion -
Training UI fix for Physicals

Keep it moving ?
2026-01-11 00:12:55 -05:00
genki
749357ba14 Label Changes - CheckPoint - Incoming Game 2026-01-10 23:29:14 -05:00
genki
52c5643b5b Added onClick from Albumviewscreen.kt 2026-01-10 22:00:23 -05:00
genki
11a1a33764 Oh yes - Thats how we do
No default params for KSP complainer fuck

UI sweeps
2026-01-10 09:44:29 -05:00
genki
f51cd4c9ba feat(training): Add parallel face detection, exclude functionality, and optimize image replacement
PERFORMANCE IMPROVEMENTS:
- Parallel face detection: 30 images now process in ~5s (was ~45s) via batched async processing
- Optimized replace: Only rescans single replaced image instead of entire set
- Memory efficient: Proper bitmap recycling in finally blocks prevents memory leaks

NEW FEATURES:
- Exclude/Include buttons: One-click removal of bad training images with instant UI feedback
- Excluded image styling: Gray overlay, disabled buttons, clear "Excluded" status
- Smart button visibility: Hide Replace/Pick Face when image excluded
- Progress tracking: Real-time callbacks during face detection scan

BUG FIXES:
- Fixed bitmap.recycle() timing to prevent corrupted face crops
- Fixed FaceDetectionHelper to recycle bitmaps only after cropping complete
- Enhanced TrainViewModel with exclude tracking and efficient state updates

UI UPDATES:
- Added ImageStatus.EXCLUDED enum value
- Updated ScanResultsScreen with exclude/include action buttons
- Enhanced color schemes for all 5 image states (Valid, Multiple, None, Error, Excluded)
- Added RemoveCircle icon for excluded images

FILES MODIFIED:
- FaceDetectionHelper.kt: Parallel processing, proper bitmap lifecycle
- TrainViewModel.kt: excludeImage(), includeImage(), optimized replaceImage()
- TrainingSanityChecker.kt: Exclusion support, progress callbacks
- ScanResultsScreen.kt: Complete exclude UI implementation

TESTING:
- 9x faster initial scan (45s → 5s for 30 images)
- 45x faster replace (45s → 1s per image)
- Instant exclude/include (<0.1s UI update)
- Minimum 15 images required for training
2026-01-10 09:44:26 -05:00
genki
52ea64f29a Oh yes - Thats how we do
No default params for KSP complainer fuck

UI sweeps
2026-01-09 19:59:44 -05:00
genki
51fdfbf3d6 Improved Training Screen and underlying
Added diagnostic view model with flag for picture detection but broke fucking everything meassing with tagDAO. au demain
2026-01-08 00:02:27 -05:00
genki
6ce115baa9 Bradeth_v1
UI improvement sweep
Underlying 'train models' backend functionality, dao and room db.
Mlmodule Hilt DI
2026-01-07 00:44:11 -05:00
genki
6734c343cc TrainScreen / FacePicker / Sanity Checking input training data (dupes, multi faces) 2026-01-02 02:20:57 -05:00
genki
22c25d5ced TODO - end of time - need to revisit anlysis results window - broke it adding the uh faePicker (needs to go in AppRoutes) 2026-01-01 01:30:08 -05:00
genki
dba64b89b6 face detection + multi faces check
filtering before crop prompt - do we need to have user crop photos with only one face?
2026-01-01 01:02:42 -05:00
genki
3f15bfabc1 Cleaner - UI ALmost and Room Photo Ingestion 2025-12-26 01:26:51 -05:00
genki
0f7f4a4201 Cleaner - Needs UI rebuild from Master TBD 2025-12-25 22:18:58 -05:00
genki
0d34a2510b Mess - Crash on boot - Backend ?? 2025-12-25 00:40:57 -05:00
genki
c458e08075 Correct schema
Meaningful queries
Proper transactional reads
2025-12-24 22:48:34 -05:00
111 changed files with 20019 additions and 519 deletions

View File

@@ -4,6 +4,14 @@
<selectionStates> <selectionStates>
<SelectionState runConfigName="app"> <SelectionState runConfigName="app">
<option name="selectionMode" value="DROPDOWN" /> <option name="selectionMode" value="DROPDOWN" />
<DropdownSelection timestamp="2026-01-08T02:44:48.809354959Z">
<Target type="DEFAULT_BOOT">
<handle>
<DeviceId pluginId="LocalEmulator" identifier="path=/home/genki/.android/avd/Medium_Phone.avd" />
</handle>
</Target>
</DropdownSelection>
<DialogSelection />
</SelectionState> </SelectionState>
</selectionStates> </selectionStates>
</component> </component>

29
.idea/deviceManager.xml generated Normal file
View File

@@ -0,0 +1,29 @@
<?xml version="1.0" encoding="UTF-8"?>
<project version="4">
<component name="DeviceTable">
<option name="columnSorters">
<list>
<ColumnSorterState>
<option name="column" value="Name" />
<option name="order" value="DESCENDING" />
</ColumnSorterState>
</list>
</option>
<option name="groupByAttributes">
<list>
<option value="Type" />
<option value="Type" />
<option value="Type" />
<option value="Type" />
<option value="Type" />
<option value="Type" />
<option value="Type" />
<option value="Type" />
<option value="Type" />
<option value="Type" />
<option value="Type" />
<option value="Type" />
</list>
</option>
</component>
</project>

View File

@@ -0,0 +1,61 @@
<component name="InspectionProjectProfileManager">
<profile version="1.0">
<option name="myName" value="Project Default" />
<inspection_tool class="ComposePreviewDimensionRespectsLimit" enabled="true" level="WARNING" enabled_by_default="true">
<option name="composableFile" value="true" />
<option name="previewFile" value="true" />
</inspection_tool>
<inspection_tool class="ComposePreviewMustBeTopLevelFunction" enabled="true" level="ERROR" enabled_by_default="true">
<option name="composableFile" value="true" />
<option name="previewFile" value="true" />
</inspection_tool>
<inspection_tool class="ComposePreviewNeedsComposableAnnotation" enabled="true" level="ERROR" enabled_by_default="true">
<option name="composableFile" value="true" />
<option name="previewFile" value="true" />
</inspection_tool>
<inspection_tool class="ComposePreviewNotSupportedInUnitTestFiles" enabled="true" level="ERROR" enabled_by_default="true">
<option name="composableFile" value="true" />
<option name="previewFile" value="true" />
</inspection_tool>
<inspection_tool class="GlancePreviewDimensionRespectsLimit" enabled="true" level="WARNING" enabled_by_default="true">
<option name="composableFile" value="true" />
</inspection_tool>
<inspection_tool class="GlancePreviewMustBeTopLevelFunction" enabled="true" level="ERROR" enabled_by_default="true">
<option name="composableFile" value="true" />
</inspection_tool>
<inspection_tool class="GlancePreviewNeedsComposableAnnotation" enabled="true" level="ERROR" enabled_by_default="true">
<option name="composableFile" value="true" />
</inspection_tool>
<inspection_tool class="GlancePreviewNotSupportedInUnitTestFiles" enabled="true" level="ERROR" enabled_by_default="true">
<option name="composableFile" value="true" />
</inspection_tool>
<inspection_tool class="PreviewAnnotationInFunctionWithParameters" enabled="true" level="ERROR" enabled_by_default="true">
<option name="composableFile" value="true" />
<option name="previewFile" value="true" />
</inspection_tool>
<inspection_tool class="PreviewApiLevelMustBeValid" enabled="true" level="ERROR" enabled_by_default="true">
<option name="composableFile" value="true" />
<option name="previewFile" value="true" />
</inspection_tool>
<inspection_tool class="PreviewDeviceShouldUseNewSpec" enabled="true" level="WEAK WARNING" enabled_by_default="true">
<option name="composableFile" value="true" />
<option name="previewFile" value="true" />
</inspection_tool>
<inspection_tool class="PreviewFontScaleMustBeGreaterThanZero" enabled="true" level="ERROR" enabled_by_default="true">
<option name="composableFile" value="true" />
<option name="previewFile" value="true" />
</inspection_tool>
<inspection_tool class="PreviewMultipleParameterProviders" enabled="true" level="ERROR" enabled_by_default="true">
<option name="composableFile" value="true" />
<option name="previewFile" value="true" />
</inspection_tool>
<inspection_tool class="PreviewParameterProviderOnFirstParameter" enabled="true" level="ERROR" enabled_by_default="true">
<option name="composableFile" value="true" />
<option name="previewFile" value="true" />
</inspection_tool>
<inspection_tool class="PreviewPickerAnnotation" enabled="true" level="ERROR" enabled_by_default="true">
<option name="composableFile" value="true" />
<option name="previewFile" value="true" />
</inspection_tool>
</profile>
</component>

View File

@@ -1,77 +1,100 @@
// build.gradle.kts (Module: :app)
plugins { plugins {
// 1. Core Android and Kotlin plugins (MUST be first) alias(libs.plugins.android.application)
id("com.android.application") alias(libs.plugins.kotlin.android)
kotlin("android") alias(libs.plugins.kotlin.compose)
alias(libs.plugins.ksp)
id("org.jetbrains.kotlin.plugin.compose") // Note: No version is specified here alias(libs.plugins.hilt.android)
} }
android { android {
// 2. Android Configuration
namespace = "com.placeholder.sherpai2" namespace = "com.placeholder.sherpai2"
compileSdk = 34 compileSdk = 35
defaultConfig { defaultConfig {
applicationId = "com.placeholder.sherpai2" applicationId = "com.placeholder.sherpai2"
minSdk = 24 minSdk = 25
targetSdk = 34 targetSdk = 35
versionCode = 1 versionCode = 1
versionName = "1.0" versionName = "1.0"
testInstrumentationRunner = "androidx.test.runner.AndroidJUnitRunner" testInstrumentationRunner = "androidx.test.runner.AndroidJUnitRunner"
} }
// 3. Kotlin & Java Settings buildTypes {
compileOptions { release {
sourceCompatibility = JavaVersion.VERSION_1_8 isMinifyEnabled = false
targetCompatibility = JavaVersion.VERSION_1_8 proguardFiles(getDefaultProguardFile("proguard-android-optimize.txt"), "proguard-rules.pro")
} }
kotlinOptions { }
jvmTarget = "1.8"
compileOptions {
sourceCompatibility = JavaVersion.VERSION_11
targetCompatibility = JavaVersion.VERSION_11
}
kotlinOptions {
jvmTarget = "11"
} }
// 4. Jetpack Compose Configuration (Crucial!)
buildFeatures { buildFeatures {
compose = true compose = true
} }
composeOptions {
kotlinCompilerExtensionVersion = "1.5.8" // Must match your Kotlin version
}
} }
dependencies { dependencies {
// --- CORE ANDROID & LIFECYCLE --- // Core & Lifecycle
implementation("androidx.core:core-ktx:1.12.0") implementation(libs.androidx.core.ktx)
implementation("androidx.lifecycle:lifecycle-runtime-compose:2.7.0") implementation(libs.androidx.lifecycle.runtime.ktx)
implementation("androidx.activity:activity-compose:1.8.2") // Fixes 'activity' ref error implementation(libs.androidx.lifecycle.viewmodel.compose)
implementation(libs.androidx.activity.compose)
// --- JETPACK COMPOSE UI (Material 3) --- // Compose
implementation("androidx.compose.ui:ui") implementation(platform(libs.androidx.compose.bom))
implementation("androidx.compose.ui:ui-graphics") implementation(libs.androidx.compose.ui)
implementation("androidx.compose.ui:ui-tooling-preview") implementation(libs.androidx.compose.ui.graphics)
implementation("androidx.compose.material3:material3") // Fixes 'material3' ref error implementation(libs.androidx.compose.ui.tooling.preview)
implementation(libs.androidx.compose.material3)
implementation(libs.androidx.compose.material.icons)
debugImplementation(libs.androidx.compose.ui.tooling)
// --- COMPOSE ICONS (Fixes 'material' and 'Icons' ref errors) --- // Hilt DI
// Uses direct string to avoid Version Catalog conflicts implementation(libs.hilt.android)
implementation("androidx.compose.material:material-icons-extended:1.6.0") ksp(libs.hilt.compiler)
implementation(libs.androidx.hilt.navigation.compose)
// --- STATE MANAGEMENT / COROUTINES --- // Navigation
implementation("androidx.lifecycle:lifecycle-viewmodel-compose:2.7.0") implementation(libs.androidx.navigation.compose)
implementation("org.jetbrains.kotlinx:kotlinx-coroutines-core:1.7.3")
implementation("org.jetbrains.kotlinx:kotlinx-coroutines-android:1.7.3")
// --- TESTING --- // Room Database
testImplementation("junit:junit:4.13.2") implementation(libs.room.runtime)
androidTestImplementation("androidx.test.ext:junit:1.1.5") implementation(libs.room.ktx)
androidTestImplementation("androidx.test.espresso:espresso-core:3.5.1") ksp(libs.room.compiler)
androidTestImplementation("androidx.compose.ui:ui-test-junit4")
debugImplementation("androidx.compose.ui:ui-tooling")
debugImplementation("androidx.compose.ui:ui-test-manifest")
implementation("androidx.compose.foundation:foundation:1.6.0") // Use your current Compose version // Coil Images
implementation("androidx.compose.material3:material3:1.2.1") // <-- Fix/Reconfirm Material 3 implementation(libs.coil.compose)
// ML Kit
implementation(libs.mlkit.face.detection)
implementation(libs.kotlinx.coroutines.play.services)
//Face Rec
implementation(libs.tensorflow.lite)
implementation(libs.tensorflow.lite.support)
// Optional: GPU acceleration
implementation(libs.tensorflow.lite.gpu)
// Gson for storing FloatArrays in Room
implementation(libs.gson)
// Zoomable
implementation(libs.zoomable)
implementation(libs.vico.compose)
implementation(libs.vico.compose.m3)
implementation(libs.vico.core)
// Workers
implementation(libs.androidx.work.runtime.ktx)
implementation(libs.androidx.hilt.work)
implementation("io.coil-kt:coil-compose:2.6.0")
} }

View File

@@ -10,7 +10,8 @@
android:label="@string/app_name" android:label="@string/app_name"
android:roundIcon="@mipmap/ic_launcher_round" android:roundIcon="@mipmap/ic_launcher_round"
android:supportsRtl="true" android:supportsRtl="true"
android:theme="@style/Theme.SherpAI2"> android:theme="@style/Theme.SherpAI2"
android:name=".SherpAIApplication">
<activity <activity
android:name=".MainActivity" android:name=".MainActivity"
android:exported="true" android:exported="true"

View File

@@ -8,63 +8,313 @@ import androidx.activity.ComponentActivity
import androidx.activity.compose.rememberLauncherForActivityResult import androidx.activity.compose.rememberLauncherForActivityResult
import androidx.activity.compose.setContent import androidx.activity.compose.setContent
import androidx.activity.result.contract.ActivityResultContracts import androidx.activity.result.contract.ActivityResultContracts
import androidx.compose.foundation.layout.Box import androidx.compose.foundation.layout.*
import androidx.compose.foundation.layout.fillMaxSize import androidx.compose.material3.*
import androidx.compose.material3.MaterialTheme import androidx.compose.runtime.*
import androidx.compose.material3.Surface
import androidx.compose.material3.Text
import androidx.compose.runtime.LaunchedEffect
import androidx.compose.runtime.getValue
import androidx.compose.runtime.mutableStateOf
import androidx.compose.runtime.remember
import androidx.compose.runtime.setValue
import androidx.compose.ui.Alignment import androidx.compose.ui.Alignment
import androidx.compose.ui.Modifier import androidx.compose.ui.Modifier
import androidx.compose.ui.text.font.FontWeight
import androidx.compose.ui.unit.dp
import androidx.core.content.ContextCompat import androidx.core.content.ContextCompat
import com.placeholder.sherpai2.presentation.MainScreen // IMPORT your main screen import androidx.lifecycle.lifecycleScope
import com.placeholder.sherpai2.domain.repository.ImageRepository
import com.placeholder.sherpai2.domain.usecase.PopulateFaceDetectionCacheUseCase
import com.placeholder.sherpai2.ui.presentation.MainScreen
import com.placeholder.sherpai2.ui.theme.SherpAI2Theme import com.placeholder.sherpai2.ui.theme.SherpAI2Theme
import dagger.hilt.android.AndroidEntryPoint
import kotlinx.coroutines.Dispatchers
import kotlinx.coroutines.launch
import kotlinx.coroutines.withContext
import javax.inject.Inject
/**
* MainActivity - TWO-PHASE STARTUP
*
* Phase 1: Image ingestion (fast - just loads URIs)
* Phase 2: Face detection cache (slower - scans for faces)
*
* App is usable immediately, both run in background.
*/
@AndroidEntryPoint
class MainActivity : ComponentActivity() { class MainActivity : ComponentActivity() {
@Inject
lateinit var imageRepository: ImageRepository
@Inject
lateinit var populateFaceCache: PopulateFaceDetectionCacheUseCase
override fun onCreate(savedInstanceState: Bundle?) { override fun onCreate(savedInstanceState: Bundle?) {
super.onCreate(savedInstanceState) super.onCreate(savedInstanceState)
// 1. Define the permission needed based on API level val storagePermission = if (Build.VERSION.SDK_INT >= Build.VERSION_CODES.TIRAMISU) {
val permission = if (Build.VERSION.SDK_INT >= Build.VERSION_CODES.TIRAMISU) {
Manifest.permission.READ_MEDIA_IMAGES Manifest.permission.READ_MEDIA_IMAGES
} else { } else {
Manifest.permission.READ_EXTERNAL_STORAGE Manifest.permission.READ_EXTERNAL_STORAGE
} }
setContent { setContent {
SherpAI2Theme { SherpAI2Theme {
// 2. State to track if permission is granted
var hasPermission by remember { var hasPermission by remember {
mutableStateOf(ContextCompat.checkSelfPermission(this, permission) == PackageManager.PERMISSION_GRANTED) mutableStateOf(
ContextCompat.checkSelfPermission(this@MainActivity, storagePermission) ==
PackageManager.PERMISSION_GRANTED
)
} }
// 3. Launcher to ask for permission var ingestionState by remember { mutableStateOf<IngestionState>(IngestionState.NotStarted) }
val launcher = rememberLauncherForActivityResult( var cacheState by remember { mutableStateOf<CacheState>(CacheState.NotStarted) }
val permissionLauncher = rememberLauncherForActivityResult(
ActivityResultContracts.RequestPermission() ActivityResultContracts.RequestPermission()
) { isGranted -> ) { granted ->
hasPermission = isGranted hasPermission = granted
} }
// 4. Trigger request on start // Phase 1: Image ingestion
LaunchedEffect(Unit) { LaunchedEffect(hasPermission) {
if (!hasPermission) launcher.launch(permission) if (hasPermission && ingestionState is IngestionState.NotStarted) {
ingestionState = IngestionState.InProgress(0, 0)
lifecycleScope.launch(Dispatchers.IO) {
try {
val existingCount = imageRepository.getImageCount()
if (existingCount > 0) {
withContext(Dispatchers.Main) {
ingestionState = IngestionState.Complete(existingCount)
}
} else {
imageRepository.ingestImagesWithProgress { current, total ->
ingestionState = IngestionState.InProgress(current, total)
}
val finalCount = imageRepository.getImageCount()
withContext(Dispatchers.Main) {
ingestionState = IngestionState.Complete(finalCount)
}
}
} catch (e: Exception) {
withContext(Dispatchers.Main) {
ingestionState = IngestionState.Error(e.message ?: "Failed to load images")
}
}
}
} else if (!hasPermission) {
permissionLauncher.launch(storagePermission)
}
} }
if (hasPermission) { // Phase 2: Face detection cache population
MainScreen() // Your existing screen that holds MainContentArea LaunchedEffect(ingestionState) {
} else { if (ingestionState is IngestionState.Complete && cacheState is CacheState.NotStarted) {
Box(modifier = Modifier.fillMaxSize(), contentAlignment = Alignment.Center) { lifecycleScope.launch(Dispatchers.IO) {
Text("Please grant storage permission to view photos.") try {
// Check if cache needs population
val stats = populateFaceCache.getCacheStats()
if (stats.needsScanning == 0) {
withContext(Dispatchers.Main) {
cacheState = CacheState.Complete(stats.imagesWithFaces, stats.imagesWithoutFaces)
}
} else {
withContext(Dispatchers.Main) {
cacheState = CacheState.InProgress(0, stats.needsScanning)
}
populateFaceCache.execute { current, total, _ ->
cacheState = CacheState.InProgress(current, total)
}
val finalStats = populateFaceCache.getCacheStats()
withContext(Dispatchers.Main) {
cacheState = CacheState.Complete(
finalStats.imagesWithFaces,
finalStats.imagesWithoutFaces
)
}
}
} catch (e: Exception) {
withContext(Dispatchers.Main) {
cacheState = CacheState.Error(e.message ?: "Failed to scan faces")
}
}
}
}
}
// UI
Box(modifier = Modifier.fillMaxSize()) {
when {
hasPermission -> {
// Main screen always visible
MainScreen()
// Progress overlays at bottom with navigation bar clearance
Column(
modifier = Modifier
.fillMaxSize()
.padding(horizontal = 16.dp)
.padding(bottom = 120.dp), // More space for nav bar + gestures
verticalArrangement = Arrangement.Bottom
) {
if (ingestionState is IngestionState.InProgress) {
IngestionProgressCard(ingestionState as IngestionState.InProgress)
Spacer(Modifier.height(8.dp))
}
if (cacheState is CacheState.InProgress) {
FaceCacheProgressCard(cacheState as CacheState.InProgress)
}
}
}
else -> {
Box(
modifier = Modifier.fillMaxSize(),
contentAlignment = Alignment.Center
) {
Column(
horizontalAlignment = Alignment.CenterHorizontally,
verticalArrangement = Arrangement.spacedBy(16.dp)
) {
Text(
"Storage Permission Required",
style = MaterialTheme.typography.titleLarge,
fontWeight = FontWeight.Bold
)
Text(
"SherpAI needs access to your photos",
style = MaterialTheme.typography.bodyMedium
)
Button(onClick = { permissionLauncher.launch(storagePermission) }) {
Text("Grant Permission")
}
}
}
}
} }
} }
} }
}
} }
}
} }
sealed class IngestionState {
object NotStarted : IngestionState()
data class InProgress(val current: Int, val total: Int) : IngestionState()
data class Complete(val imageCount: Int) : IngestionState()
data class Error(val message: String) : IngestionState()
}
sealed class CacheState {
object NotStarted : CacheState()
data class InProgress(val current: Int, val total: Int) : CacheState()
data class Complete(val withFaces: Int, val withoutFaces: Int) : CacheState()
data class Error(val message: String) : CacheState()
}
@Composable
fun IngestionProgressCard(state: IngestionState.InProgress) {
Card(
modifier = Modifier.fillMaxWidth(),
colors = CardDefaults.cardColors(
containerColor = MaterialTheme.colorScheme.primaryContainer
),
elevation = CardDefaults.cardElevation(defaultElevation = 8.dp)
) {
Column(
modifier = Modifier
.fillMaxWidth()
.padding(16.dp),
verticalArrangement = Arrangement.spacedBy(12.dp)
) {
Row(
modifier = Modifier.fillMaxWidth(),
horizontalArrangement = Arrangement.SpaceBetween,
verticalAlignment = Alignment.CenterVertically
) {
Text(
text = "Loading photos...",
style = MaterialTheme.typography.titleMedium,
fontWeight = FontWeight.Bold
)
if (state.total > 0) {
Text(
text = "${state.current} / ${state.total}",
style = MaterialTheme.typography.bodyMedium,
color = MaterialTheme.colorScheme.primary
)
}
}
if (state.total > 0) {
LinearProgressIndicator(
progress = { state.current.toFloat() / state.total.toFloat() },
modifier = Modifier.fillMaxWidth(),
)
} else {
LinearProgressIndicator(modifier = Modifier.fillMaxWidth())
}
Text(
text = "You can use the app while photos load",
style = MaterialTheme.typography.bodySmall,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
}
}
}
@Composable
fun FaceCacheProgressCard(state: CacheState.InProgress) {
Card(
modifier = Modifier.fillMaxWidth(),
colors = CardDefaults.cardColors(
containerColor = MaterialTheme.colorScheme.secondaryContainer
),
elevation = CardDefaults.cardElevation(defaultElevation = 8.dp)
) {
Column(
modifier = Modifier
.fillMaxWidth()
.padding(16.dp),
verticalArrangement = Arrangement.spacedBy(12.dp)
) {
Row(
modifier = Modifier.fillMaxWidth(),
horizontalArrangement = Arrangement.SpaceBetween,
verticalAlignment = Alignment.CenterVertically
) {
Text(
text = "Scanning for faces...",
style = MaterialTheme.typography.titleMedium,
fontWeight = FontWeight.Bold
)
if (state.total > 0) {
Text(
text = "${state.current} / ${state.total}",
style = MaterialTheme.typography.bodyMedium,
color = MaterialTheme.colorScheme.secondary
)
}
}
if (state.total > 0) {
LinearProgressIndicator(
progress = { state.current.toFloat() / state.total.toFloat() },
modifier = Modifier.fillMaxWidth(),
)
} else {
LinearProgressIndicator(modifier = Modifier.fillMaxWidth())
}
Text(
text = "Face filters will work once scanning completes",
style = MaterialTheme.typography.bodySmall,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
}
}
}

View File

@@ -0,0 +1,24 @@
package com.placeholder.sherpai2
import android.app.Application
import androidx.hilt.work.HiltWorkerFactory
import androidx.work.Configuration
import dagger.hilt.android.HiltAndroidApp
import javax.inject.Inject
/**
* SherpAIApplication - ENHANCED with WorkManager support
*
* Now supports background cache population via Hilt Workers
*/
@HiltAndroidApp
class SherpAIApplication : Application(), Configuration.Provider {
@Inject
lateinit var workerFactory: HiltWorkerFactory
override val workManagerConfiguration: Configuration
get() = Configuration.Builder()
.setWorkerFactory(workerFactory)
.build()
}

View File

@@ -0,0 +1,166 @@
package com.placeholder.sherpai2.data.local
import androidx.room.Database
import androidx.room.RoomDatabase
import androidx.sqlite.db.SupportSQLiteDatabase
import androidx.room.migration.Migration
import com.placeholder.sherpai2.data.local.dao.*
import com.placeholder.sherpai2.data.local.entity.*
/**
* AppDatabase - Complete database for SherpAI2
*
* VERSION 8 - PHASE 2: Multi-centroid face models + age tagging
* - Added PersonEntity.isChild, siblingIds, familyGroupId
* - Changed FaceModelEntity.embedding → centroidsJson (multi-centroid)
* - Added PersonAgeTagEntity table for searchable age tags
*
* MIGRATION STRATEGY:
* - Development: fallbackToDestructiveMigration (fresh install)
* - Production: Add MIGRATION_7_8 before release
*/
@Database(
entities = [
// ===== CORE ENTITIES =====
ImageEntity::class,
TagEntity::class,
EventEntity::class,
ImageTagEntity::class,
ImageEventEntity::class,
// ===== FACE RECOGNITION =====
PersonEntity::class,
FaceModelEntity::class,
PhotoFaceTagEntity::class,
PersonAgeTagEntity::class, // NEW: Age tagging
// ===== COLLECTIONS =====
CollectionEntity::class,
CollectionImageEntity::class,
CollectionFilterEntity::class
],
version = 8, // INCREMENTED for Phase 2
exportSchema = false
)
abstract class AppDatabase : RoomDatabase() {
// ===== CORE DAOs =====
abstract fun imageDao(): ImageDao
abstract fun tagDao(): TagDao
abstract fun eventDao(): EventDao
abstract fun imageTagDao(): ImageTagDao
abstract fun imageEventDao(): ImageEventDao
abstract fun imageAggregateDao(): ImageAggregateDao
// ===== FACE RECOGNITION DAOs =====
abstract fun personDao(): PersonDao
abstract fun faceModelDao(): FaceModelDao
abstract fun photoFaceTagDao(): PhotoFaceTagDao
abstract fun personAgeTagDao(): PersonAgeTagDao // NEW
// ===== COLLECTIONS DAO =====
abstract fun collectionDao(): CollectionDao
}
/**
* MIGRATION 7 → 8 (Phase 2)
*
* Changes:
* 1. Add isChild, siblingIds, familyGroupId to persons table
* 2. Rename embedding → centroidsJson in face_models table
* 3. Create person_age_tags table
*/
val MIGRATION_7_8 = object : Migration(7, 8) {
override fun migrate(database: SupportSQLiteDatabase) {
// ===== STEP 1: Update persons table =====
database.execSQL("ALTER TABLE persons ADD COLUMN isChild INTEGER NOT NULL DEFAULT 0")
database.execSQL("ALTER TABLE persons ADD COLUMN siblingIds TEXT DEFAULT NULL")
database.execSQL("ALTER TABLE persons ADD COLUMN familyGroupId TEXT DEFAULT NULL")
// Create index on familyGroupId for sibling queries
database.execSQL("CREATE INDEX IF NOT EXISTS index_persons_familyGroupId ON persons(familyGroupId)")
// ===== STEP 2: Update face_models table =====
// Rename embedding column to centroidsJson
// SQLite doesn't support RENAME COLUMN directly, so we need to:
// 1. Create new table with new schema
// 2. Copy data (converting single embedding to centroid JSON)
// 3. Drop old table
// 4. Rename new table
// Create new table
database.execSQL("""
CREATE TABLE IF NOT EXISTS face_models_new (
id TEXT PRIMARY KEY NOT NULL,
personId TEXT NOT NULL,
centroidsJson TEXT NOT NULL,
trainingImageCount INTEGER NOT NULL,
averageConfidence REAL NOT NULL,
createdAt INTEGER NOT NULL,
updatedAt INTEGER NOT NULL,
lastUsed INTEGER,
isActive INTEGER NOT NULL,
FOREIGN KEY(personId) REFERENCES persons(id) ON DELETE CASCADE
)
""")
// Copy data, converting embedding to centroidsJson format
// This converts single embedding to a list with one centroid
database.execSQL("""
INSERT INTO face_models_new
SELECT
id,
personId,
'[{"embedding":' || REPLACE(REPLACE(embedding, ',', ','), ',', ',') || ',"effectiveTimestamp":' || createdAt || ',"ageAtCapture":null,"photoCount":' || trainingImageCount || ',"timeRangeMonths":12,"avgConfidence":' || averageConfidence || '}]' as centroidsJson,
trainingImageCount,
averageConfidence,
createdAt,
updatedAt,
lastUsed,
isActive
FROM face_models
""")
// Drop old table
database.execSQL("DROP TABLE face_models")
// Rename new table
database.execSQL("ALTER TABLE face_models_new RENAME TO face_models")
// Recreate index
database.execSQL("CREATE UNIQUE INDEX IF NOT EXISTS index_face_models_personId ON face_models(personId)")
// ===== STEP 3: Create person_age_tags table =====
database.execSQL("""
CREATE TABLE IF NOT EXISTS person_age_tags (
id TEXT PRIMARY KEY NOT NULL,
personId TEXT NOT NULL,
imageId TEXT NOT NULL,
ageAtCapture INTEGER NOT NULL,
tagValue TEXT NOT NULL,
confidence REAL NOT NULL,
createdAt INTEGER NOT NULL,
FOREIGN KEY(personId) REFERENCES persons(id) ON DELETE CASCADE,
FOREIGN KEY(imageId) REFERENCES images(imageId) ON DELETE CASCADE
)
""")
// Create indices for fast lookups
database.execSQL("CREATE INDEX IF NOT EXISTS index_person_age_tags_personId ON person_age_tags(personId)")
database.execSQL("CREATE INDEX IF NOT EXISTS index_person_age_tags_imageId ON person_age_tags(imageId)")
database.execSQL("CREATE INDEX IF NOT EXISTS index_person_age_tags_ageAtCapture ON person_age_tags(ageAtCapture)")
database.execSQL("CREATE INDEX IF NOT EXISTS index_person_age_tags_tagValue ON person_age_tags(tagValue)")
}
}
/**
* PRODUCTION MIGRATION NOTES:
*
* Before shipping to users, update DatabaseModule to use migration:
*
* Room.databaseBuilder(context, AppDatabase::class.java, "sherpai.db")
* .addMigrations(MIGRATION_7_8) // Add this
* // .fallbackToDestructiveMigration() // Remove this
* .build()
*/

View File

@@ -0,0 +1,216 @@
package com.placeholder.sherpai2.data.local.dao
import androidx.room.*
import com.placeholder.sherpai2.data.local.entity.*
import com.placeholder.sherpai2.data.local.model.CollectionWithDetails
import kotlinx.coroutines.flow.Flow
/**
* CollectionDao - Manage user collections
*/
@Dao
interface CollectionDao {
// ==========================================
// BASIC OPERATIONS
// ==========================================
@Insert(onConflict = OnConflictStrategy.REPLACE)
suspend fun insert(collection: CollectionEntity): Long
@Update
suspend fun update(collection: CollectionEntity)
@Delete
suspend fun delete(collection: CollectionEntity)
@Query("DELETE FROM collections WHERE collectionId = :collectionId")
suspend fun deleteById(collectionId: String)
@Query("SELECT * FROM collections WHERE collectionId = :collectionId")
suspend fun getById(collectionId: String): CollectionEntity?
@Query("SELECT * FROM collections WHERE collectionId = :collectionId")
fun getByIdFlow(collectionId: String): Flow<CollectionEntity?>
// ==========================================
// LIST QUERIES
// ==========================================
/**
* Get all collections ordered by pinned, then by creation date
*/
@Query("""
SELECT * FROM collections
ORDER BY isPinned DESC, createdAt DESC
""")
fun getAllCollections(): Flow<List<CollectionEntity>>
@Query("""
SELECT * FROM collections
WHERE type = :type
ORDER BY isPinned DESC, createdAt DESC
""")
fun getCollectionsByType(type: String): Flow<List<CollectionEntity>>
@Query("SELECT * FROM collections WHERE type = 'FAVORITE' LIMIT 1")
suspend fun getFavoriteCollection(): CollectionEntity?
// ==========================================
// COLLECTION WITH DETAILS
// ==========================================
/**
* Get collection with actual photo count
*/
@Transaction
@Query("""
SELECT
c.*,
(SELECT COUNT(*)
FROM collection_images ci
WHERE ci.collectionId = c.collectionId) as actualPhotoCount
FROM collections c
WHERE c.collectionId = :collectionId
""")
fun getCollectionWithDetails(collectionId: String): Flow<CollectionWithDetails?>
// ==========================================
// IMAGE MANAGEMENT
// ==========================================
@Insert(onConflict = OnConflictStrategy.REPLACE)
suspend fun addImage(collectionImage: CollectionImageEntity)
@Insert(onConflict = OnConflictStrategy.REPLACE)
suspend fun addImages(collectionImages: List<CollectionImageEntity>)
@Query("""
DELETE FROM collection_images
WHERE collectionId = :collectionId AND imageId = :imageId
""")
suspend fun removeImage(collectionId: String, imageId: String)
@Query("DELETE FROM collection_images WHERE collectionId = :collectionId")
suspend fun clearAllImages(collectionId: String)
@Query("""
SELECT i.* FROM images i
JOIN collection_images ci ON i.imageId = ci.imageId
WHERE ci.collectionId = :collectionId
ORDER BY ci.sortOrder ASC, ci.addedAt DESC
""")
fun getImagesInCollection(collectionId: String): Flow<List<ImageEntity>>
@Query("""
SELECT i.* FROM images i
JOIN collection_images ci ON i.imageId = ci.imageId
WHERE ci.collectionId = :collectionId
ORDER BY ci.sortOrder ASC, ci.addedAt DESC
LIMIT 4
""")
suspend fun getPreviewImages(collectionId: String): List<ImageEntity>
@Query("""
SELECT COUNT(*) FROM collection_images
WHERE collectionId = :collectionId
""")
suspend fun getPhotoCount(collectionId: String): Int
@Query("""
SELECT EXISTS(
SELECT 1 FROM collection_images
WHERE collectionId = :collectionId AND imageId = :imageId
)
""")
suspend fun containsImage(collectionId: String, imageId: String): Boolean
// ==========================================
// FILTER MANAGEMENT (for SMART collections)
// ==========================================
@Insert(onConflict = OnConflictStrategy.REPLACE)
suspend fun insertFilter(filter: CollectionFilterEntity)
@Insert(onConflict = OnConflictStrategy.REPLACE)
suspend fun insertFilters(filters: List<CollectionFilterEntity>)
@Query("DELETE FROM collection_filters WHERE collectionId = :collectionId")
suspend fun clearFilters(collectionId: String)
@Query("""
SELECT * FROM collection_filters
WHERE collectionId = :collectionId
ORDER BY createdAt ASC
""")
suspend fun getFilters(collectionId: String): List<CollectionFilterEntity>
@Query("""
SELECT * FROM collection_filters
WHERE collectionId = :collectionId
ORDER BY createdAt ASC
""")
fun getFiltersFlow(collectionId: String): Flow<List<CollectionFilterEntity>>
// ==========================================
// STATISTICS
// ==========================================
@Query("SELECT COUNT(*) FROM collections")
suspend fun getCollectionCount(): Int
@Query("SELECT COUNT(*) FROM collections WHERE type = 'SMART'")
suspend fun getSmartCollectionCount(): Int
@Query("SELECT COUNT(*) FROM collections WHERE type = 'STATIC'")
suspend fun getStaticCollectionCount(): Int
@Query("""
SELECT SUM(photoCount) FROM collections
""")
suspend fun getTotalPhotosInCollections(): Int?
// ==========================================
// UPDATES
// ==========================================
/**
* Update photo count cache (call after adding/removing images)
*/
@Query("""
UPDATE collections
SET photoCount = (
SELECT COUNT(*) FROM collection_images
WHERE collectionId = :collectionId
),
updatedAt = :updatedAt
WHERE collectionId = :collectionId
""")
suspend fun updatePhotoCount(collectionId: String, updatedAt: Long)
@Query("""
UPDATE collections
SET coverImageUri = :imageUri, updatedAt = :updatedAt
WHERE collectionId = :collectionId
""")
suspend fun updateCoverImage(collectionId: String, imageUri: String?, updatedAt: Long)
@Query("""
UPDATE collections
SET isPinned = :isPinned, updatedAt = :updatedAt
WHERE collectionId = :collectionId
""")
suspend fun updatePinned(collectionId: String, isPinned: Boolean, updatedAt: Long)
@Query("""
UPDATE collections
SET name = :name, description = :description, updatedAt = :updatedAt
WHERE collectionId = :collectionId
""")
suspend fun updateDetails(
collectionId: String,
name: String,
description: String?,
updatedAt: Long
)
}

View File

@@ -0,0 +1,26 @@
package com.placeholder.sherpai2.data.local.dao
import androidx.room.Dao
import androidx.room.Insert
import androidx.room.OnConflictStrategy
import androidx.room.Query
import com.placeholder.sherpai2.data.local.entity.EventEntity
@Dao
interface EventDao {
@Insert(onConflict = OnConflictStrategy.REPLACE)
suspend fun insert(event: EventEntity)
/**
* Find events covering a timestamp.
*
* This is the backbone of auto-tagging by date.
*/
@Query("""
SELECT * FROM events
WHERE :timestamp BETWEEN startDate AND endDate
AND isHidden = 0
""")
suspend fun findEventsForTimestamp(timestamp: Long): List<EventEntity>
}

View File

@@ -0,0 +1,44 @@
package com.placeholder.sherpai2.data.local.dao
import androidx.room.*
import kotlinx.coroutines.flow.Flow
import com.placeholder.sherpai2.data.local.entity.FaceModelEntity
/**
* FaceModelDao - Manages face recognition models
*
* PRIMARY KEY TYPE: String (UUID)
* FOREIGN KEY: personId (String)
*/
@Dao
interface FaceModelDao {
@Insert(onConflict = OnConflictStrategy.REPLACE)
suspend fun insertFaceModel(faceModel: FaceModelEntity): Long // Row ID
@Update
suspend fun updateFaceModel(faceModel: FaceModelEntity)
@Query("UPDATE face_models SET lastUsed = :timestamp WHERE id = :faceModelId")
suspend fun updateLastUsed(faceModelId: String, timestamp: Long)
@Query("SELECT * FROM face_models WHERE id = :faceModelId")
suspend fun getFaceModelById(faceModelId: String): FaceModelEntity?
@Query("SELECT * FROM face_models WHERE personId = :personId AND isActive = 1")
suspend fun getFaceModelByPersonId(personId: String): FaceModelEntity?
@Query("SELECT * FROM face_models WHERE isActive = 1 ORDER BY lastUsed DESC")
suspend fun getAllActiveFaceModels(): List<FaceModelEntity>
@Query("SELECT * FROM face_models WHERE isActive = 1 ORDER BY lastUsed DESC")
fun getAllActiveFaceModelsFlow(): Flow<List<FaceModelEntity>>
@Query("DELETE FROM face_models WHERE id = :faceModelId")
suspend fun deleteFaceModelById(faceModelId: String)
@Query("UPDATE face_models SET isActive = 0 WHERE id = :faceModelId")
suspend fun deactivateFaceModel(faceModelId: String)
@Query("SELECT COUNT(*) FROM face_models WHERE isActive = 1")
suspend fun getActiveFaceModelCount(): Int
}

View File

@@ -0,0 +1,48 @@
package com.placeholder.sherpai2.data.local.dao
import androidx.room.Dao
import androidx.room.Query
import androidx.room.Transaction
import com.placeholder.sherpai2.data.local.model.ImageWithEverything
import kotlinx.coroutines.flow.Flow
@Dao
interface ImageAggregateDao {
/**
* Observe a fully-hydrated image object.
*/
@Transaction
@Query("""
SELECT * FROM images
WHERE imageId = :imageId
""")
fun observeImageWithEverything(
imageId: String
): Flow<ImageWithEverything>
/**
* Observe all images.
*/
@Transaction
@Query("""
SELECT * FROM images
ORDER BY capturedAt DESC
""")
fun observeAllImagesWithEverything(): Flow<List<ImageWithEverything>>
/**
* Observe images filtered by tag value.
*
* Joins images -> image_tags -> tags
*/
@Transaction
@Query("""
SELECT images.* FROM images
INNER JOIN image_tags ON images.imageId = image_tags.imageId
INNER JOIN tags ON tags.tagId = image_tags.tagId
WHERE tags.value = :tag
ORDER BY images.capturedAt DESC
""")
fun observeImagesWithTag(tag: String): Flow<List<ImageWithEverything>>
}

View File

@@ -0,0 +1,440 @@
package com.placeholder.sherpai2.data.local.dao
import androidx.room.Dao
import androidx.room.Insert
import androidx.room.OnConflictStrategy
import androidx.room.Query
import androidx.room.Transaction
import com.placeholder.sherpai2.data.local.entity.ImageEntity
import com.placeholder.sherpai2.data.local.model.ImageWithEverything
import kotlinx.coroutines.flow.Flow
/**
* Data classes for statistics queries
*/
data class DateCount(
val date: String, // YYYY-MM-DD format
val count: Int
)
data class MonthCount(
val month: String, // YYYY-MM format
val count: Int
)
data class YearCount(
val year: String, // YYYY format
val count: Int
)
data class DayOfWeekCount(
val dayOfWeek: Int, // 0 = Sunday, 6 = Saturday
val count: Int
)
data class HourCount(
val hour: Int, // 0-23
val count: Int
)
/**
* Face detection cache statistics
*/
data class FaceCacheStats(
val totalImages: Int,
val imagesWithFaceCache: Int,
val imagesWithFaces: Int,
val imagesWithoutFaces: Int,
val needsScanning: Int
)
@Dao
interface ImageDao {
/**
* Insert images.
*
* IGNORE prevents duplicate insertion
* when sha256 or imageUri already exists.
*/
@Insert(onConflict = OnConflictStrategy.IGNORE)
suspend fun insertImages(images: List<ImageEntity>)
/**
* Get image by ID.
*/
@Query("SELECT * FROM images WHERE imageId = :imageId")
suspend fun getImageById(imageId: String): ImageEntity?
/**
* Stream images ordered by capture time (newest first).
*
* Flow is critical:
* - UI auto-updates
* - No manual refresh
*/
@Query("""
SELECT * FROM images
ORDER BY capturedAt DESC
""")
fun observeAllImages(): Flow<List<ImageEntity>>
/**
* Fetch images in a time range.
* Used for:
* - event auto-assignment
* - timeline views
*/
@Query("""
SELECT * FROM images
WHERE capturedAt BETWEEN :start AND :end
ORDER BY capturedAt ASC
""")
suspend fun getImagesInRange(
start: Long,
end: Long
): List<ImageEntity>
@Transaction
@Query("SELECT * FROM images ORDER BY capturedAt DESC LIMIT :limit")
fun getRecentImages(limit: Int): Flow<List<ImageWithEverything>>
@Query("SELECT COUNT(*) > 0 FROM images WHERE sha256 = :sha256")
suspend fun existsBySha256(sha256: String): Boolean
@Insert(onConflict = OnConflictStrategy.IGNORE)
suspend fun insert(image: ImageEntity)
/**
* Get images by list of IDs.
*/
@Query("SELECT * FROM images WHERE imageId IN (:imageIds)")
suspend fun getImagesByIds(imageIds: List<String>): List<ImageEntity>
@Query("SELECT COUNT(*) FROM images")
suspend fun getImageCount(): Int
/**
* Get all images (for utilities processing)
*/
@Query("SELECT * FROM images ORDER BY capturedAt DESC")
suspend fun getAllImages(): List<ImageEntity>
/**
* Get all images sorted by time (for burst detection)
*/
@Query("SELECT * FROM images ORDER BY capturedAt ASC")
suspend fun getAllImagesSortedByTime(): List<ImageEntity>
// ==========================================
// FACE DETECTION CACHE QUERIES - CRITICAL FOR OPTIMIZATION
// ==========================================
/**
* Get all images that have faces (cached).
* This is the PRIMARY optimization query.
*
* Use this for person scanning instead of scanning ALL images.
* Estimated speed improvement: 50-70% for typical photo libraries.
*/
@Query("""
SELECT * FROM images
WHERE hasFaces = 1
AND faceDetectionVersion = :currentVersion
ORDER BY capturedAt DESC
""")
suspend fun getImagesWithFaces(currentVersion: Int = ImageEntity.CURRENT_FACE_DETECTION_VERSION): List<ImageEntity>
/**
* Get images with faces, limited (for progressive scanning)
*/
@Query("""
SELECT * FROM images
WHERE hasFaces = 1
AND faceDetectionVersion = :currentVersion
ORDER BY capturedAt DESC
LIMIT :limit
""")
suspend fun getImagesWithFacesLimited(
limit: Int,
currentVersion: Int = ImageEntity.CURRENT_FACE_DETECTION_VERSION
): List<ImageEntity>
/**
* Get images with a specific face count.
* Use cases:
* - Solo photos (faceCount = 1)
* - Couple photos (faceCount = 2)
* - Filter out groups (faceCount <= 2)
*/
@Query("""
SELECT * FROM images
WHERE hasFaces = 1
AND faceCount = :count
AND faceDetectionVersion = :currentVersion
ORDER BY capturedAt DESC
""")
suspend fun getImagesByFaceCount(
count: Int,
currentVersion: Int = ImageEntity.CURRENT_FACE_DETECTION_VERSION
): List<ImageEntity>
/**
* Get images with face count in range.
* Examples:
* - Solo or couple: minFaces=1, maxFaces=2
* - Groups only: minFaces=3, maxFaces=999
*/
@Query("""
SELECT * FROM images
WHERE hasFaces = 1
AND faceCount BETWEEN :minFaces AND :maxFaces
AND faceDetectionVersion = :currentVersion
ORDER BY capturedAt DESC
""")
suspend fun getImagesByFaceCountRange(
minFaces: Int,
maxFaces: Int,
currentVersion: Int = ImageEntity.CURRENT_FACE_DETECTION_VERSION
): List<ImageEntity>
/**
* Get images that need face detection scanning.
* These images have:
* - Never been scanned (hasFaces = null)
* - Old detection version
* - Invalid cache
*/
@Query("""
SELECT * FROM images
WHERE hasFaces IS NULL
OR faceDetectionVersion IS NULL
OR faceDetectionVersion < :currentVersion
ORDER BY capturedAt DESC
""")
suspend fun getImagesNeedingFaceDetection(
currentVersion: Int = ImageEntity.CURRENT_FACE_DETECTION_VERSION
): List<ImageEntity>
/**
* Get count of images needing face detection.
*/
@Query("""
SELECT COUNT(*) FROM images
WHERE hasFaces IS NULL
OR faceDetectionVersion IS NULL
OR faceDetectionVersion < :currentVersion
""")
suspend fun getImagesNeedingFaceDetectionCount(
currentVersion: Int = ImageEntity.CURRENT_FACE_DETECTION_VERSION
): Int
/**
* Update face detection cache for a single image.
* Called after detecting faces in an image.
*/
@Query("""
UPDATE images
SET hasFaces = :hasFaces,
faceCount = :faceCount,
facesLastDetected = :timestamp,
faceDetectionVersion = :version
WHERE imageId = :imageId
""")
suspend fun updateFaceDetectionCache(
imageId: String,
hasFaces: Boolean,
faceCount: Int,
timestamp: Long = System.currentTimeMillis(),
version: Int = ImageEntity.CURRENT_FACE_DETECTION_VERSION
)
/**
* Batch update face detection cache.
* More efficient when updating many images at once.
*
* Note: Room doesn't support batch updates directly,
* so this needs to be called multiple times in a transaction.
*/
@Transaction
suspend fun updateFaceDetectionCacheBatch(updates: List<FaceDetectionCacheUpdate>) {
updates.forEach { update ->
updateFaceDetectionCache(
imageId = update.imageId,
hasFaces = update.hasFaces,
faceCount = update.faceCount,
timestamp = update.timestamp,
version = update.version
)
}
}
/**
* Get face detection cache statistics.
* Useful for UI display and determining background scan needs.
*/
@Query("""
SELECT
COUNT(*) as totalImages,
SUM(CASE WHEN hasFaces IS NOT NULL THEN 1 ELSE 0 END) as imagesWithFaceCache,
SUM(CASE WHEN hasFaces = 1 THEN 1 ELSE 0 END) as imagesWithFaces,
SUM(CASE WHEN hasFaces = 0 THEN 1 ELSE 0 END) as imagesWithoutFaces,
SUM(CASE WHEN hasFaces IS NULL OR faceDetectionVersion < :currentVersion THEN 1 ELSE 0 END) as needsScanning
FROM images
""")
suspend fun getFaceCacheStats(
currentVersion: Int = ImageEntity.CURRENT_FACE_DETECTION_VERSION
): FaceCacheStats?
/**
* Invalidate face detection cache (force re-scan).
* Call this when upgrading face detection algorithm.
*/
@Query("""
UPDATE images
SET faceDetectionVersion = NULL
WHERE faceDetectionVersion < :newVersion
""")
suspend fun invalidateFaceDetectionCache(newVersion: Int)
// ==========================================
// STATISTICS QUERIES
// ==========================================
/**
* Get photo counts by date (daily granularity)
* Returns all days that have at least one photo
*/
@Query("""
SELECT
date(capturedAt/1000, 'unixepoch') as date,
COUNT(*) as count
FROM images
GROUP BY date
ORDER BY date ASC
""")
suspend fun getPhotoCountsByDate(): List<DateCount>
/**
* Get photo counts by month (monthly granularity)
*/
@Query("""
SELECT
strftime('%Y-%m', capturedAt/1000, 'unixepoch') as month,
COUNT(*) as count
FROM images
GROUP BY month
ORDER BY month ASC
""")
suspend fun getPhotoCountsByMonth(): List<MonthCount>
/**
* Get photo counts by year (yearly granularity)
*/
@Query("""
SELECT
strftime('%Y', capturedAt/1000, 'unixepoch') as year,
COUNT(*) as count
FROM images
GROUP BY year
ORDER BY year DESC
""")
suspend fun getPhotoCountsByYear(): List<YearCount>
/**
* Get photo counts by year (Flow version for reactive UI)
*/
@Query("""
SELECT
strftime('%Y', capturedAt/1000, 'unixepoch') as year,
COUNT(*) as count
FROM images
GROUP BY year
ORDER BY year DESC
""")
fun getPhotoCountsByYearFlow(): Flow<List<YearCount>>
/**
* Get photo counts by day of week (0 = Sunday, 6 = Saturday)
* Shows which days you take the most photos
*/
@Query("""
SELECT
CAST(strftime('%w', capturedAt/1000, 'unixepoch') AS INTEGER) as dayOfWeek,
COUNT(*) as count
FROM images
GROUP BY dayOfWeek
ORDER BY dayOfWeek ASC
""")
suspend fun getPhotoCountsByDayOfWeek(): List<DayOfWeekCount>
/**
* Get photo counts by hour of day (0-23)
* Shows when you take the most photos
*/
@Query("""
SELECT
CAST(strftime('%H', capturedAt/1000, 'unixepoch') AS INTEGER) as hour,
COUNT(*) as count
FROM images
GROUP BY hour
ORDER BY hour ASC
""")
suspend fun getPhotoCountsByHour(): List<HourCount>
/**
* Get earliest and latest photo timestamps
* Used for date range calculations
*/
@Query("""
SELECT
MIN(capturedAt) as earliest,
MAX(capturedAt) as latest
FROM images
""")
suspend fun getPhotoDateRange(): PhotoDateRange?
/**
* Get photo count for a specific year
*/
@Query("""
SELECT COUNT(*) FROM images
WHERE strftime('%Y', capturedAt/1000, 'unixepoch') = :year
""")
suspend fun getPhotoCountForYear(year: String): Int
/**
* Get average photos per day (for stats display)
*/
@Query("""
SELECT
CAST(COUNT(*) AS REAL) /
CAST((MAX(capturedAt) - MIN(capturedAt)) / 86400000 AS REAL) as avgPerDay
FROM images
WHERE (SELECT COUNT(*) FROM images) > 0
""")
suspend fun getAveragePhotosPerDay(): Float?
@Query("SELECT * FROM images WHERE hasFaces = 1 ORDER BY faceCount DESC")
suspend fun getImagesWithFaces(): List<ImageEntity>
}
/**
* Data class for date range result
*/
data class PhotoDateRange(
val earliest: Long,
val latest: Long
)
/**
* Data class for batch face detection cache updates
*/
data class FaceDetectionCacheUpdate(
val imageId: String,
val hasFaces: Boolean,
val faceCount: Int,
val timestamp: Long = System.currentTimeMillis(),
val version: Int = ImageEntity.CURRENT_FACE_DETECTION_VERSION
)

View File

@@ -0,0 +1,23 @@
package com.placeholder.sherpai2.data.local.dao
import androidx.room.Dao
import androidx.room.Insert
import androidx.room.OnConflictStrategy
import androidx.room.Query
import com.placeholder.sherpai2.data.local.entity.ImageEventEntity
@Dao
interface ImageEventDao {
@Insert(onConflict = OnConflictStrategy.REPLACE)
suspend fun upsert(entity: ImageEventEntity)
/**
* Images associated with an event.
*/
@Query("""
SELECT imageId FROM image_events
WHERE eventId = :eventId
""")
suspend fun findImagesForEvent(eventId: String): List<String>
}

View File

@@ -0,0 +1,142 @@
package com.placeholder.sherpai2.data.local.dao
import androidx.room.Dao
import androidx.room.Insert
import androidx.room.OnConflictStrategy
import androidx.room.Query
import androidx.room.Transaction
import com.placeholder.sherpai2.data.local.entity.ImageTagEntity
import com.placeholder.sherpai2.data.local.entity.TagEntity
import kotlinx.coroutines.flow.Flow
/**
* Data class for burst statistics
*/
data class BurstStats(
val totalBurstPhotos: Int,
val estimatedBurstGroups: Int,
val burstRepresentatives: Int
)
@Dao
interface ImageTagDao {
@Insert(onConflict = OnConflictStrategy.REPLACE)
suspend fun upsert(imageTag: ImageTagEntity)
@Query("""
SELECT * FROM image_tags
WHERE imageId = :imageId
AND visibility != 'HIDDEN'
""")
fun observeTagsForImage(imageId: String): Flow<List<ImageTagEntity>>
/**
* FIXED: Removed default parameter
*/
@Query("""
SELECT imageId FROM image_tags
WHERE tagId = :tagId
AND visibility = 'PUBLIC'
AND confidence >= :minConfidence
""")
suspend fun findImagesByTag(
tagId: String,
minConfidence: Float
): List<String>
@Transaction
@Query("""
SELECT t.*
FROM tags t
INNER JOIN image_tags it ON t.tagId = it.tagId
WHERE it.imageId = :imageId AND it.visibility = 'PUBLIC'
""")
fun getTagsForImage(imageId: String): Flow<List<TagEntity>>
/**
* Insert image tag (for utilities tagging)
*/
@Insert(onConflict = OnConflictStrategy.IGNORE)
suspend fun insert(imageTag: ImageTagEntity): Long
// ==========================================
// BURST STATISTICS - ADDED FOR STATS SECTION
// ==========================================
/**
* Get comprehensive burst statistics
* Returns total burst photos, estimated groups, and representative count
*/
@Query("""
SELECT
(SELECT COUNT(DISTINCT it.imageId)
FROM image_tags it
INNER JOIN tags t ON it.tagId = t.tagId
WHERE t.value = 'burst') as totalBurstPhotos,
(SELECT COUNT(DISTINCT it.imageId) / 3
FROM image_tags it
INNER JOIN tags t ON it.tagId = t.tagId
WHERE t.value = 'burst') as estimatedBurstGroups,
(SELECT COUNT(DISTINCT it.imageId)
FROM image_tags it
INNER JOIN tags t ON it.tagId = t.tagId
WHERE t.value = 'burst_representative') as burstRepresentatives
""")
suspend fun getBurstStats(): BurstStats?
/**
* Get burst statistics (Flow version for reactive UI)
*/
@Query("""
SELECT
(SELECT COUNT(DISTINCT it.imageId)
FROM image_tags it
INNER JOIN tags t ON it.tagId = t.tagId
WHERE t.value = 'burst') as totalBurstPhotos,
(SELECT COUNT(DISTINCT it.imageId) / 3
FROM image_tags it
INNER JOIN tags t ON it.tagId = t.tagId
WHERE t.value = 'burst') as estimatedBurstGroups,
(SELECT COUNT(DISTINCT it.imageId)
FROM image_tags it
INNER JOIN tags t ON it.tagId = t.tagId
WHERE t.value = 'burst_representative') as burstRepresentatives
""")
fun getBurstStatsFlow(): Flow<BurstStats?>
/**
* Get count of burst photos
*/
@Query("""
SELECT COUNT(DISTINCT it.imageId)
FROM image_tags it
INNER JOIN tags t ON it.tagId = t.tagId
WHERE t.value = 'burst'
""")
suspend fun getBurstPhotoCount(): Int
/**
* Get count of burst representative photos
* (photos marked as the best in each burst sequence)
*/
@Query("""
SELECT COUNT(DISTINCT it.imageId)
FROM image_tags it
INNER JOIN tags t ON it.tagId = t.tagId
WHERE t.value = 'burst_representative'
""")
suspend fun getBurstRepresentativeCount(): Int
/**
* Get estimated number of burst groups
* Assumes average of 3 photos per burst
*/
@Query("""
SELECT COUNT(DISTINCT it.imageId) / 3
FROM image_tags it
INNER JOIN tags t ON it.tagId = t.tagId
WHERE t.value = 'burst'
""")
suspend fun getEstimatedBurstGroupCount(): Int
}

View File

@@ -0,0 +1,51 @@
package com.placeholder.sherpai2.data.local.dao
import androidx.room.*
import com.placeholder.sherpai2.data.local.entity.PersonEntity
import kotlinx.coroutines.flow.Flow
@Dao
interface PersonDao {
@Insert(onConflict = OnConflictStrategy.REPLACE)
suspend fun insert(person: PersonEntity): Long
@Insert(onConflict = OnConflictStrategy.REPLACE)
suspend fun insertAll(persons: List<PersonEntity>)
@Update
suspend fun update(person: PersonEntity)
/**
* FIXED: Removed default parameter
*/
@Query("UPDATE persons SET updatedAt = :timestamp WHERE id = :personId")
suspend fun updateTimestamp(personId: String, timestamp: Long)
@Delete
suspend fun delete(person: PersonEntity)
@Query("DELETE FROM persons WHERE id = :personId")
suspend fun deleteById(personId: String)
@Query("SELECT * FROM persons WHERE id = :personId")
suspend fun getPersonById(personId: String): PersonEntity?
@Query("SELECT * FROM persons WHERE id IN (:personIds)")
suspend fun getPersonsByIds(personIds: List<String>): List<PersonEntity>
@Query("SELECT * FROM persons ORDER BY name ASC")
suspend fun getAllPersons(): List<PersonEntity>
@Query("SELECT * FROM persons ORDER BY name ASC")
fun getAllPersonsFlow(): Flow<List<PersonEntity>>
@Query("SELECT * FROM persons WHERE name LIKE '%' || :query || '%' ORDER BY name ASC")
suspend fun searchByName(query: String): List<PersonEntity>
@Query("SELECT COUNT(*) FROM persons")
suspend fun getPersonCount(): Int
@Query("SELECT EXISTS(SELECT 1 FROM persons WHERE id = :personId)")
suspend fun personExists(personId: String): Boolean
}

View File

@@ -0,0 +1,104 @@
package com.placeholder.sherpai2.data.local.dao
import androidx.room.*
import com.placeholder.sherpai2.data.local.entity.PersonAgeTagEntity
import kotlinx.coroutines.flow.Flow
/**
* PersonAgeTagDao - Manage searchable age tags for children
*
* USAGE EXAMPLES:
* - Search "emma age 3" → getImageIdsForTag("emma_age3")
* - Find all photos of Emma at age 5 → getImageIdsForPersonAtAge(emmaId, 5)
* - Get age progression → getTagsForPerson(emmaId) sorted by age
*/
@Dao
interface PersonAgeTagDao {
@Insert(onConflict = OnConflictStrategy.REPLACE)
suspend fun insertTag(tag: PersonAgeTagEntity)
@Insert(onConflict = OnConflictStrategy.REPLACE)
suspend fun insertTags(tags: List<PersonAgeTagEntity>)
/**
* Get all age tags for a person (sorted by age)
* Useful for age progression timeline
*/
@Query("SELECT * FROM person_age_tags WHERE personId = :personId ORDER BY ageAtCapture ASC")
suspend fun getTagsForPerson(personId: String): List<PersonAgeTagEntity>
/**
* Get all age tags for an image
*/
@Query("SELECT * FROM person_age_tags WHERE imageId = :imageId")
suspend fun getTagsForImage(imageId: String): List<PersonAgeTagEntity>
/**
* Search by tag value (e.g., "emma_age3")
* Returns all image IDs matching this tag
*/
@Query("SELECT DISTINCT imageId FROM person_age_tags WHERE tagValue = :tagValue")
suspend fun getImageIdsForTag(tagValue: String): List<String>
/**
* Get images of a person at a specific age
*/
@Query("SELECT DISTINCT imageId FROM person_age_tags WHERE personId = :personId AND ageAtCapture = :age")
suspend fun getImageIdsForPersonAtAge(personId: String, age: Int): List<String>
/**
* Get images of a person in an age range
*/
@Query("""
SELECT DISTINCT imageId FROM person_age_tags
WHERE personId = :personId
AND ageAtCapture BETWEEN :minAge AND :maxAge
ORDER BY ageAtCapture ASC
""")
suspend fun getImageIdsForPersonAgeRange(personId: String, minAge: Int, maxAge: Int): List<String>
/**
* Get all unique ages for a person (for age picker UI)
*/
@Query("SELECT DISTINCT ageAtCapture FROM person_age_tags WHERE personId = :personId ORDER BY ageAtCapture ASC")
suspend fun getAgesForPerson(personId: String): List<Int>
/**
* Delete all tags for a person
*/
@Query("DELETE FROM person_age_tags WHERE personId = :personId")
suspend fun deleteTagsForPerson(personId: String)
/**
* Delete all tags for an image
*/
@Query("DELETE FROM person_age_tags WHERE imageId = :imageId")
suspend fun deleteTagsForImage(imageId: String)
/**
* Get count of photos at each age (for statistics)
*/
@Query("""
SELECT ageAtCapture, COUNT(DISTINCT imageId) as count
FROM person_age_tags
WHERE personId = :personId
GROUP BY ageAtCapture
ORDER BY ageAtCapture ASC
""")
suspend fun getPhotoCountByAge(personId: String): List<AgePhotoCount>
/**
* Flow version for reactive UI
*/
@Query("SELECT * FROM person_age_tags WHERE personId = :personId ORDER BY ageAtCapture ASC")
fun getTagsForPersonFlow(personId: String): Flow<List<PersonAgeTagEntity>>
}
/**
* Data class for age photo count statistics
*/
data class AgePhotoCount(
val ageAtCapture: Int,
val count: Int
)

View File

@@ -0,0 +1,91 @@
package com.placeholder.sherpai2.data.local.dao
import androidx.room.*
import kotlinx.coroutines.flow.Flow
import com.placeholder.sherpai2.data.local.entity.PhotoFaceTagEntity
@Dao
interface PhotoFaceTagDao {
@Insert(onConflict = OnConflictStrategy.REPLACE)
suspend fun insertTag(tag: PhotoFaceTagEntity): Long
@Insert(onConflict = OnConflictStrategy.REPLACE)
suspend fun insertTags(tags: List<PhotoFaceTagEntity>)
@Update
suspend fun updateTag(tag: PhotoFaceTagEntity)
/**
* FIXED: Removed default parameter
*/
@Query("UPDATE photo_face_tags SET verifiedByUser = 1, verifiedAt = :timestamp WHERE id = :tagId")
suspend fun markTagAsVerified(tagId: String, timestamp: Long)
// ===== QUERY BY IMAGE =====
@Query("SELECT * FROM photo_face_tags WHERE imageId = :imageId")
suspend fun getTagsForImage(imageId: String): List<PhotoFaceTagEntity>
@Query("SELECT COUNT(*) FROM photo_face_tags WHERE imageId = :imageId")
suspend fun getFaceCountForImage(imageId: String): Int
@Query("SELECT EXISTS(SELECT 1 FROM photo_face_tags WHERE imageId = :imageId AND faceModelId = :faceModelId)")
suspend fun imageHasPerson(imageId: String, faceModelId: String): Boolean
// ===== QUERY BY FACE MODEL =====
@Query("SELECT DISTINCT imageId FROM photo_face_tags WHERE faceModelId = :faceModelId ORDER BY detectedAt DESC")
suspend fun getImageIdsForFaceModel(faceModelId: String): List<String>
@Query("SELECT DISTINCT imageId FROM photo_face_tags WHERE faceModelId = :faceModelId ORDER BY detectedAt DESC")
fun getImageIdsForFaceModelFlow(faceModelId: String): Flow<List<String>>
@Query("SELECT faceModelId, COUNT(DISTINCT imageId) as photoCount FROM photo_face_tags GROUP BY faceModelId")
suspend fun getPhotoCountPerFaceModel(): List<FaceModelPhotoCount>
@Query("SELECT * FROM photo_face_tags WHERE faceModelId = :faceModelId ORDER BY detectedAt DESC")
suspend fun getAllTagsForFaceModel(faceModelId: String): List<PhotoFaceTagEntity>
// ===== DELETE =====
@Delete
suspend fun deleteTag(tag: PhotoFaceTagEntity)
@Query("DELETE FROM photo_face_tags WHERE id = :tagId")
suspend fun deleteTagById(tagId: String)
@Query("DELETE FROM photo_face_tags WHERE faceModelId = :faceModelId")
suspend fun deleteTagsForFaceModel(faceModelId: String)
@Query("DELETE FROM photo_face_tags WHERE imageId = :imageId")
suspend fun deleteTagsForImage(imageId: String)
// ===== STATISTICS =====
/**
* FIXED: Removed default parameter
*/
@Query("SELECT * FROM photo_face_tags WHERE confidence < :threshold ORDER BY confidence ASC")
suspend fun getLowConfidenceTags(threshold: Float): List<PhotoFaceTagEntity>
@Query("SELECT * FROM photo_face_tags WHERE verifiedByUser = 0 ORDER BY detectedAt DESC")
suspend fun getUnverifiedTags(): List<PhotoFaceTagEntity>
@Query("SELECT COUNT(*) FROM photo_face_tags WHERE verifiedByUser = 0")
suspend fun getUnverifiedTagCount(): Int
@Query("SELECT AVG(confidence) FROM photo_face_tags WHERE faceModelId = :faceModelId")
suspend fun getAverageConfidenceForFaceModel(faceModelId: String): Float?
/**
* FIXED: Removed default parameter
*/
@Query("SELECT * FROM photo_face_tags ORDER BY detectedAt DESC LIMIT :limit")
suspend fun getRecentlyDetectedFaces(limit: Int): List<PhotoFaceTagEntity>
}
data class FaceModelPhotoCount(
val faceModelId: String,
val photoCount: Int
)

View File

@@ -0,0 +1,297 @@
package com.placeholder.sherpai2.data.local.dao
import androidx.room.Dao
import androidx.room.Insert
import androidx.room.OnConflictStrategy
import androidx.room.Query
import com.placeholder.sherpai2.data.local.entity.ImageEntity
import com.placeholder.sherpai2.data.local.entity.TagEntity
import com.placeholder.sherpai2.data.local.entity.TagWithUsage
import kotlinx.coroutines.flow.Flow
/**
* Data class for tag statistics
*/
data class TagStat(
val tagValue: String,
val tagType: String,
val imageCount: Int,
val tagId: String
)
/**
* TagDao - Tag management with face recognition integration
*
* NO DEFAULT PARAMETERS - Room doesn't support them in @Query methods
*/
@Dao
interface TagDao {
// ======================
// BASIC OPERATIONS
// ======================
@Insert(onConflict = OnConflictStrategy.IGNORE)
suspend fun insert(tag: TagEntity): Long
@Query("SELECT * FROM tags WHERE value = :value LIMIT 1")
suspend fun getByValue(value: String): TagEntity?
@Query("SELECT * FROM tags WHERE tagId = :tagId")
suspend fun getById(tagId: String): TagEntity?
@Query("SELECT * FROM tags ORDER BY value ASC")
suspend fun getAll(): List<TagEntity>
@Query("SELECT * FROM tags ORDER BY value ASC")
fun getAllFlow(): Flow<List<TagEntity>>
@Query("SELECT * FROM tags WHERE type = :type ORDER BY value ASC")
suspend fun getByType(type: String): List<TagEntity>
@Query("DELETE FROM tags WHERE tagId = :tagId")
suspend fun delete(tagId: String)
// ======================
// STATISTICS (returns TagWithUsage)
// ======================
/**
* Get most used tags WITH usage counts
*
* @param limit Maximum number of tags to return
*/
@Query("""
SELECT t.tagId, t.type, t.value, t.createdAt,
COUNT(it.imageId) as usage_count
FROM tags t
LEFT JOIN image_tags it ON t.tagId = it.tagId
GROUP BY t.tagId
ORDER BY usage_count DESC
LIMIT :limit
""")
suspend fun getMostUsedTags(limit: Int): List<TagWithUsage>
/**
* Get tag usage count
*/
@Query("""
SELECT COUNT(DISTINCT it.imageId)
FROM image_tags it
WHERE it.tagId = :tagId
""")
suspend fun getTagUsageCount(tagId: String): Int
// ======================
// PERSON INTEGRATION
// ======================
/**
* Get all tags used for images containing a specific person
*/
@Query("""
SELECT DISTINCT t.* FROM tags t
INNER JOIN image_tags it ON t.tagId = it.tagId
INNER JOIN photo_face_tags pft ON it.imageId = pft.imageId
INNER JOIN face_models fm ON pft.faceModelId = fm.id
WHERE fm.personId = :personId
ORDER BY t.value ASC
""")
suspend fun getTagsForPerson(personId: String): List<TagEntity>
/**
* Get images that have both a specific tag AND contain a specific person
*/
@Query("""
SELECT DISTINCT i.* FROM images i
INNER JOIN image_tags it ON i.imageId = it.imageId
INNER JOIN photo_face_tags pft ON i.imageId = pft.imageId
INNER JOIN face_models fm ON pft.faceModelId = fm.id
WHERE it.tagId = :tagId AND fm.personId = :personId
ORDER BY i.capturedAt DESC
""")
suspend fun getImagesWithTagAndPerson(
tagId: String,
personId: String
): List<ImageEntity>
/**
* Get images with tag and person as Flow
*/
@Query("""
SELECT DISTINCT i.* FROM images i
INNER JOIN image_tags it ON i.imageId = it.imageId
INNER JOIN photo_face_tags pft ON i.imageId = pft.imageId
INNER JOIN face_models fm ON pft.faceModelId = fm.id
WHERE it.tagId = :tagId AND fm.personId = :personId
ORDER BY i.capturedAt DESC
""")
fun getImagesWithTagAndPersonFlow(
tagId: String,
personId: String
): Flow<List<ImageEntity>>
/**
* Count images with tag and person
*/
@Query("""
SELECT COUNT(DISTINCT i.imageId) FROM images i
INNER JOIN image_tags it ON i.imageId = it.imageId
INNER JOIN photo_face_tags pft ON i.imageId = pft.imageId
INNER JOIN face_models fm ON pft.faceModelId = fm.id
WHERE it.tagId = :tagId AND fm.personId = :personId
""")
suspend fun countImagesWithTagAndPerson(
tagId: String,
personId: String
): Int
// ======================
// AUTO-SUGGESTIONS
// ======================
/**
* Suggest tags based on person's relationship
*
* @param limit Maximum number of suggestions
*/
@Query("""
SELECT DISTINCT t.* FROM tags t
INNER JOIN image_tags it ON t.tagId = it.tagId
INNER JOIN photo_face_tags pft ON it.imageId = pft.imageId
INNER JOIN face_models fm ON pft.faceModelId = fm.id
INNER JOIN persons p ON fm.personId = p.id
WHERE p.relationship = :relationship
AND p.id != :excludePersonId
GROUP BY t.tagId
ORDER BY COUNT(it.imageId) DESC
LIMIT :limit
""")
suspend fun suggestTagsBasedOnRelationship(
relationship: String,
excludePersonId: String,
limit: Int
): List<TagEntity>
/**
* Get tags commonly used with this tag
*
* @param limit Maximum number of related tags
*/
@Query("""
SELECT DISTINCT t2.* FROM tags t2
INNER JOIN image_tags it2 ON t2.tagId = it2.tagId
WHERE it2.imageId IN (
SELECT it1.imageId FROM image_tags it1
WHERE it1.tagId = :tagId
)
AND t2.tagId != :tagId
GROUP BY t2.tagId
ORDER BY COUNT(it2.imageId) DESC
LIMIT :limit
""")
suspend fun getRelatedTags(
tagId: String,
limit: Int
): List<TagEntity>
// ======================
// SEARCH
// ======================
/**
* Search tags by value (partial match)
*
* @param limit Maximum number of results
*/
@Query("""
SELECT * FROM tags
WHERE value LIKE '%' || :query || '%'
ORDER BY value ASC
LIMIT :limit
""")
suspend fun searchTags(query: String, limit: Int): List<TagEntity>
/**
* Search tags with usage count
*
* @param limit Maximum number of results
*/
@Query("""
SELECT t.tagId, t.type, t.value, t.createdAt,
COUNT(it.imageId) as usage_count
FROM tags t
LEFT JOIN image_tags it ON t.tagId = it.tagId
WHERE t.value LIKE '%' || :query || '%'
GROUP BY t.tagId
ORDER BY usage_count DESC, t.value ASC
LIMIT :limit
""")
suspend fun searchTagsWithUsage(query: String, limit: Int): List<TagWithUsage>
// ==========================================
// STATISTICS QUERIES - ADDED FOR STATS SECTION
// ==========================================
/**
* Get system tag statistics (for utilities stats display)
* Returns tag value, type, and count of tagged images
*/
@Query("""
SELECT
t.value as tagValue,
t.type as tagType,
COUNT(DISTINCT it.imageId) as imageCount,
t.tagId as tagId
FROM tags t
INNER JOIN image_tags it ON t.tagId = it.tagId
WHERE t.type = 'SYSTEM'
GROUP BY t.tagId
ORDER BY imageCount DESC
""")
suspend fun getSystemTagStats(): List<TagStat>
/**
* Get system tag statistics (Flow version for reactive UI)
*/
@Query("""
SELECT
t.value as tagValue,
t.type as tagType,
COUNT(DISTINCT it.imageId) as imageCount,
t.tagId as tagId
FROM tags t
INNER JOIN image_tags it ON t.tagId = it.tagId
WHERE t.type = 'SYSTEM'
GROUP BY t.tagId
ORDER BY imageCount DESC
""")
fun getSystemTagStatsFlow(): Flow<List<TagStat>>
/**
* Get count of photos with a specific system tag
*/
@Query("""
SELECT COUNT(DISTINCT it.imageId)
FROM image_tags it
INNER JOIN tags t ON it.tagId = t.tagId
WHERE t.value = :tagValue AND t.type = 'SYSTEM'
""")
suspend fun getSystemTagCount(tagValue: String): Int
/**
* Get all tag types with counts
* Shows breakdown of SYSTEM vs USER vs GENERIC tags
*/
@Query("""
SELECT
t.type as tagValue,
t.type as tagType,
COUNT(DISTINCT t.tagId) as imageCount,
'' as tagId
FROM tags t
GROUP BY t.type
ORDER BY imageCount DESC
""")
suspend fun getTagTypeBreakdown(): List<TagStat>
}

View File

@@ -0,0 +1,107 @@
package com.placeholder.sherpai2.data.local.entity
import androidx.room.Entity
import androidx.room.Index
import androidx.room.PrimaryKey
import java.util.UUID
/**
* CollectionEntity - User-created photo collections
*
* Types:
* - SMART: Dynamic collection based on filters (re-evaluated)
* - STATIC: Fixed snapshot of photos
* - FAVORITE: Special favorites collection
*/
@Entity(
tableName = "collections",
indices = [
Index(value = ["name"]),
Index(value = ["type"]),
Index(value = ["createdAt"])
]
)
data class CollectionEntity(
@PrimaryKey
val collectionId: String,
val name: String,
val description: String?,
/**
* Cover image (auto-selected or user-chosen)
*/
val coverImageUri: String?,
/**
* SMART | STATIC | FAVORITE
*/
val type: String,
/**
* Cached photo count for performance
*/
val photoCount: Int,
val createdAt: Long,
val updatedAt: Long,
/**
* Pinned to top of collections list
*/
val isPinned: Boolean
) {
companion object {
fun createSmart(
name: String,
description: String? = null
): CollectionEntity {
val now = System.currentTimeMillis()
return CollectionEntity(
collectionId = UUID.randomUUID().toString(),
name = name,
description = description,
coverImageUri = null,
type = "SMART",
photoCount = 0,
createdAt = now,
updatedAt = now,
isPinned = false
)
}
fun createStatic(
name: String,
description: String? = null,
photoCount: Int = 0
): CollectionEntity {
val now = System.currentTimeMillis()
return CollectionEntity(
collectionId = UUID.randomUUID().toString(),
name = name,
description = description,
coverImageUri = null,
type = "STATIC",
photoCount = photoCount,
createdAt = now,
updatedAt = now,
isPinned = false
)
}
fun createFavorite(): CollectionEntity {
val now = System.currentTimeMillis()
return CollectionEntity(
collectionId = "favorites",
name = "Favorites",
description = "Your favorite photos",
coverImageUri = null,
type = "FAVORITE",
photoCount = 0,
createdAt = now,
updatedAt = now,
isPinned = true
)
}
}
}

View File

@@ -0,0 +1,70 @@
package com.placeholder.sherpai2.data.local.entity
import androidx.room.Entity
import androidx.room.ForeignKey
import androidx.room.Index
import androidx.room.PrimaryKey
import java.util.UUID
/**
* CollectionFilterEntity - Filters for SMART collections
*
* Filter Types:
* - PERSON_INCLUDE: Person must be in photo
* - PERSON_EXCLUDE: Person must NOT be in photo
* - TAG_INCLUDE: Tag must be present
* - TAG_EXCLUDE: Tag must NOT be present
* - DATE_RANGE: Date filter (TODAY, THIS_WEEK, etc)
*/
@Entity(
tableName = "collection_filters",
foreignKeys = [
ForeignKey(
entity = CollectionEntity::class,
parentColumns = ["collectionId"],
childColumns = ["collectionId"],
onDelete = ForeignKey.CASCADE
)
],
indices = [
Index("collectionId"),
Index("filterType")
]
)
data class CollectionFilterEntity(
@PrimaryKey
val filterId: String,
val collectionId: String,
/**
* PERSON_INCLUDE | PERSON_EXCLUDE | TAG_INCLUDE | TAG_EXCLUDE | DATE_RANGE
*/
val filterType: String,
/**
* The filter value:
* - For PERSON_*: personId
* - For TAG_*: tag value
* - For DATE_RANGE: "TODAY", "THIS_WEEK", etc
*/
val filterValue: String,
val createdAt: Long
) {
companion object {
fun create(
collectionId: String,
filterType: String,
filterValue: String
): CollectionFilterEntity {
return CollectionFilterEntity(
filterId = UUID.randomUUID().toString(),
collectionId = collectionId,
filterType = filterType,
filterValue = filterValue,
createdAt = System.currentTimeMillis()
)
}
}
}

View File

@@ -0,0 +1,50 @@
package com.placeholder.sherpai2.data.local.entity
import androidx.room.Entity
import androidx.room.ForeignKey
import androidx.room.Index
/**
* CollectionImageEntity - Join table linking collections to images
*
* Supports:
* - Custom sort order
* - Timestamp when added
*/
@Entity(
tableName = "collection_images",
primaryKeys = ["collectionId", "imageId"],
foreignKeys = [
ForeignKey(
entity = CollectionEntity::class,
parentColumns = ["collectionId"],
childColumns = ["collectionId"],
onDelete = ForeignKey.CASCADE
),
ForeignKey(
entity = ImageEntity::class,
parentColumns = ["imageId"],
childColumns = ["imageId"],
onDelete = ForeignKey.CASCADE
)
],
indices = [
Index("collectionId"),
Index("imageId"),
Index("addedAt")
]
)
data class CollectionImageEntity(
val collectionId: String,
val imageId: String,
/**
* When this image was added to the collection
*/
val addedAt: Long,
/**
* Custom sort order (lower = earlier)
*/
val sortOrder: Int
)

View File

@@ -0,0 +1,44 @@
package com.placeholder.sherpai2.data.local.entity
import androidx.room.Entity
import androidx.room.Index
import androidx.room.PrimaryKey
/**
* Represents a meaningful event spanning a time range.
*
* Events allow auto-association of images by timestamp.
*/
@Entity(
tableName = "events",
indices = [
Index(value = ["startDate"]),
Index(value = ["endDate"])
]
)
data class EventEntity(
@PrimaryKey
val eventId: String,
val name: String,
/**
* Inclusive start date (UTC millis).
*/
val startDate: Long,
/**
* Inclusive end date (UTC millis).
*/
val endDate: Long,
val location: String?,
/**
* 0.0 1.0 importance weight
*/
val importance: Float,
val isHidden: Boolean
)

View File

@@ -0,0 +1,455 @@
package com.placeholder.sherpai2.data.local.entity
import androidx.room.ColumnInfo
import androidx.room.Entity
import androidx.room.ForeignKey
import androidx.room.Index
import androidx.room.PrimaryKey
import org.json.JSONArray
import org.json.JSONObject
import java.util.UUID
/**
* PersonEntity - ENHANCED with child tracking and sibling relationships
*/
@Entity(
tableName = "persons",
indices = [
Index(value = ["name"]),
Index(value = ["familyGroupId"])
]
)
data class PersonEntity(
@PrimaryKey
@ColumnInfo(name = "id")
val id: String,
@ColumnInfo(name = "name")
val name: String,
@ColumnInfo(name = "dateOfBirth")
val dateOfBirth: Long?,
@ColumnInfo(name = "isChild")
val isChild: Boolean, // NEW: Auto-set based on age
@ColumnInfo(name = "siblingIds")
val siblingIds: String?, // NEW: JSON list ["uuid1", "uuid2"]
@ColumnInfo(name = "familyGroupId")
val familyGroupId: String?, // NEW: UUID for family unit
@ColumnInfo(name = "relationship")
val relationship: String?,
@ColumnInfo(name = "createdAt")
val createdAt: Long,
@ColumnInfo(name = "updatedAt")
val updatedAt: Long
) {
companion object {
fun create(
name: String,
dateOfBirth: Long? = null,
isChild: Boolean = false,
siblingIds: List<String> = emptyList(),
relationship: String? = null
): PersonEntity {
val now = System.currentTimeMillis()
// Create family group if siblings exist
val familyGroupId = if (siblingIds.isNotEmpty()) {
UUID.randomUUID().toString()
} else null
return PersonEntity(
id = UUID.randomUUID().toString(),
name = name,
dateOfBirth = dateOfBirth,
isChild = isChild,
siblingIds = if (siblingIds.isNotEmpty()) {
JSONArray(siblingIds).toString()
} else null,
familyGroupId = familyGroupId,
relationship = relationship,
createdAt = now,
updatedAt = now
)
}
}
fun getSiblingIds(): List<String> {
return if (siblingIds != null) {
try {
val jsonArray = JSONArray(siblingIds)
(0 until jsonArray.length()).map { jsonArray.getString(it) }
} catch (e: Exception) {
emptyList()
}
} else emptyList()
}
fun getAge(): Int? {
if (dateOfBirth == null) return null
val now = System.currentTimeMillis()
val ageInMillis = now - dateOfBirth
return (ageInMillis / (1000L * 60 * 60 * 24 * 365)).toInt()
}
fun getRelationshipEmoji(): String {
return when (relationship) {
"Family" -> "👨‍👩‍👧‍👦"
"Friend" -> "🤝"
"Partner" -> "❤️"
"Child" -> "👶"
"Parent" -> "👪"
"Sibling" -> "👫"
"Colleague" -> "💼"
else -> "👤"
}
}
}
/**
* FaceModelEntity - MULTI-CENTROID support for temporal tracking
*/
@Entity(
tableName = "face_models",
foreignKeys = [
ForeignKey(
entity = PersonEntity::class,
parentColumns = ["id"],
childColumns = ["personId"],
onDelete = ForeignKey.CASCADE
)
],
indices = [Index(value = ["personId"], unique = true)]
)
data class FaceModelEntity(
@PrimaryKey
@ColumnInfo(name = "id")
val id: String,
@ColumnInfo(name = "personId")
val personId: String,
@ColumnInfo(name = "centroidsJson")
val centroidsJson: String, // NEW: List<TemporalCentroid> as JSON
@ColumnInfo(name = "trainingImageCount")
val trainingImageCount: Int,
@ColumnInfo(name = "averageConfidence")
val averageConfidence: Float,
@ColumnInfo(name = "createdAt")
val createdAt: Long,
@ColumnInfo(name = "updatedAt")
val updatedAt: Long,
@ColumnInfo(name = "lastUsed")
val lastUsed: Long?,
@ColumnInfo(name = "isActive")
val isActive: Boolean
) {
companion object {
/**
* Backwards compatible create() method
* Used by existing FaceRecognitionRepository code
*/
fun create(
personId: String,
embeddingArray: FloatArray,
trainingImageCount: Int,
averageConfidence: Float
): FaceModelEntity {
return createFromEmbedding(personId, embeddingArray, trainingImageCount, averageConfidence)
}
/**
* Create from single embedding (backwards compatible)
*/
fun createFromEmbedding(
personId: String,
embeddingArray: FloatArray,
trainingImageCount: Int,
averageConfidence: Float
): FaceModelEntity {
val now = System.currentTimeMillis()
val centroid = TemporalCentroid(
embedding = embeddingArray.toList(),
effectiveTimestamp = now,
ageAtCapture = null,
photoCount = trainingImageCount,
timeRangeMonths = 12,
avgConfidence = averageConfidence
)
return FaceModelEntity(
id = UUID.randomUUID().toString(),
personId = personId,
centroidsJson = serializeCentroids(listOf(centroid)),
trainingImageCount = trainingImageCount,
averageConfidence = averageConfidence,
createdAt = now,
updatedAt = now,
lastUsed = null,
isActive = true
)
}
/**
* Create from multiple centroids (temporal tracking)
*/
fun createFromCentroids(
personId: String,
centroids: List<TemporalCentroid>,
trainingImageCount: Int,
averageConfidence: Float
): FaceModelEntity {
val now = System.currentTimeMillis()
return FaceModelEntity(
id = UUID.randomUUID().toString(),
personId = personId,
centroidsJson = serializeCentroids(centroids),
trainingImageCount = trainingImageCount,
averageConfidence = averageConfidence,
createdAt = now,
updatedAt = now,
lastUsed = null,
isActive = true
)
}
/**
* Serialize list of centroids to JSON
*/
private fun serializeCentroids(centroids: List<TemporalCentroid>): String {
val jsonArray = JSONArray()
centroids.forEach { centroid ->
val jsonObj = JSONObject()
jsonObj.put("embedding", JSONArray(centroid.embedding))
jsonObj.put("effectiveTimestamp", centroid.effectiveTimestamp)
jsonObj.put("ageAtCapture", centroid.ageAtCapture)
jsonObj.put("photoCount", centroid.photoCount)
jsonObj.put("timeRangeMonths", centroid.timeRangeMonths)
jsonObj.put("avgConfidence", centroid.avgConfidence)
jsonArray.put(jsonObj)
}
return jsonArray.toString()
}
/**
* Deserialize JSON to list of centroids
*/
private fun deserializeCentroids(json: String): List<TemporalCentroid> {
val jsonArray = JSONArray(json)
return (0 until jsonArray.length()).map { i ->
val jsonObj = jsonArray.getJSONObject(i)
val embeddingArray = jsonObj.getJSONArray("embedding")
val embedding = (0 until embeddingArray.length()).map { j ->
embeddingArray.getDouble(j).toFloat()
}
TemporalCentroid(
embedding = embedding,
effectiveTimestamp = jsonObj.getLong("effectiveTimestamp"),
ageAtCapture = if (jsonObj.isNull("ageAtCapture")) null else jsonObj.getDouble("ageAtCapture").toFloat(),
photoCount = jsonObj.getInt("photoCount"),
timeRangeMonths = jsonObj.getInt("timeRangeMonths"),
avgConfidence = jsonObj.getDouble("avgConfidence").toFloat()
)
}
}
}
fun getCentroids(): List<TemporalCentroid> {
return try {
FaceModelEntity.deserializeCentroids(centroidsJson)
} catch (e: Exception) {
emptyList()
}
}
// Backwards compatibility: get first centroid as single embedding
fun getEmbeddingArray(): FloatArray {
val centroids = getCentroids()
return if (centroids.isNotEmpty()) {
centroids.first().getEmbeddingArray()
} else {
FloatArray(192) // Empty embedding
}
}
}
/**
* TemporalCentroid - Represents a face appearance at a specific time period
*/
data class TemporalCentroid(
val embedding: List<Float>, // 192D vector
val effectiveTimestamp: Long, // Center of time window
val ageAtCapture: Float?, // Age in years (for children)
val photoCount: Int, // Number of photos in this cluster
val timeRangeMonths: Int, // Width of time window (e.g., 6 months)
val avgConfidence: Float // Quality indicator
) {
fun getEmbeddingArray(): FloatArray = embedding.toFloatArray()
}
/**
* PhotoFaceTagEntity - Unchanged
*/
@Entity(
tableName = "photo_face_tags",
foreignKeys = [
ForeignKey(
entity = ImageEntity::class,
parentColumns = ["imageId"],
childColumns = ["imageId"],
onDelete = ForeignKey.CASCADE
),
ForeignKey(
entity = FaceModelEntity::class,
parentColumns = ["id"],
childColumns = ["faceModelId"],
onDelete = ForeignKey.CASCADE
)
],
indices = [
Index(value = ["imageId"]),
Index(value = ["faceModelId"]),
Index(value = ["imageId", "faceModelId"])
]
)
data class PhotoFaceTagEntity(
@PrimaryKey
@ColumnInfo(name = "id")
val id: String,
@ColumnInfo(name = "imageId")
val imageId: String,
@ColumnInfo(name = "faceModelId")
val faceModelId: String,
@ColumnInfo(name = "boundingBox")
val boundingBox: String,
@ColumnInfo(name = "confidence")
val confidence: Float,
@ColumnInfo(name = "embedding")
val embedding: String,
@ColumnInfo(name = "detectedAt")
val detectedAt: Long,
@ColumnInfo(name = "verifiedByUser")
val verifiedByUser: Boolean,
@ColumnInfo(name = "verifiedAt")
val verifiedAt: Long?
) {
companion object {
fun create(
imageId: String,
faceModelId: String,
boundingBox: android.graphics.Rect,
confidence: Float,
faceEmbedding: FloatArray
): PhotoFaceTagEntity {
return PhotoFaceTagEntity(
id = UUID.randomUUID().toString(),
imageId = imageId,
faceModelId = faceModelId,
boundingBox = "${boundingBox.left},${boundingBox.top},${boundingBox.right},${boundingBox.bottom}",
confidence = confidence,
embedding = faceEmbedding.joinToString(","),
detectedAt = System.currentTimeMillis(),
verifiedByUser = false,
verifiedAt = null
)
}
}
fun getBoundingBox(): android.graphics.Rect {
val parts = boundingBox.split(",").map { it.toInt() }
return android.graphics.Rect(parts[0], parts[1], parts[2], parts[3])
}
fun getEmbeddingArray(): FloatArray {
return embedding.split(",").map { it.toFloat() }.toFloatArray()
}
}
/**
* PersonAgeTagEntity - NEW: Searchable age tags
*/
@Entity(
tableName = "person_age_tags",
foreignKeys = [
ForeignKey(
entity = PersonEntity::class,
parentColumns = ["id"],
childColumns = ["personId"],
onDelete = ForeignKey.CASCADE
),
ForeignKey(
entity = ImageEntity::class,
parentColumns = ["imageId"],
childColumns = ["imageId"],
onDelete = ForeignKey.CASCADE
)
],
indices = [
Index(value = ["personId"]),
Index(value = ["imageId"]),
Index(value = ["ageAtCapture"]),
Index(value = ["tagValue"])
]
)
data class PersonAgeTagEntity(
@PrimaryKey
@ColumnInfo(name = "id")
val id: String,
@ColumnInfo(name = "personId")
val personId: String,
@ColumnInfo(name = "imageId")
val imageId: String,
@ColumnInfo(name = "ageAtCapture")
val ageAtCapture: Int,
@ColumnInfo(name = "tagValue")
val tagValue: String, // e.g., "emma_age3"
@ColumnInfo(name = "confidence")
val confidence: Float,
@ColumnInfo(name = "createdAt")
val createdAt: Long
) {
companion object {
fun create(
personId: String,
personName: String,
imageId: String,
ageAtCapture: Int,
confidence: Float
): PersonAgeTagEntity {
return PersonAgeTagEntity(
id = UUID.randomUUID().toString(),
personId = personId,
imageId = imageId,
ageAtCapture = ageAtCapture,
tagValue = "${personName.lowercase().replace(" ", "_")}_age$ageAtCapture",
confidence = confidence,
createdAt = System.currentTimeMillis()
)
}
}
}

View File

@@ -0,0 +1,175 @@
package com.placeholder.sherpai2.data.local.entity
import androidx.room.Entity
import androidx.room.Index
import androidx.room.PrimaryKey
/**
* Represents a single image on the device.
*
* This entity is intentionally immutable (mostly):
* - imageUri identifies where the image lives
* - sha256 prevents duplicates
* - capturedAt is the EXIF timestamp
*
* FACE DETECTION CACHE (mutable for performance):
* - hasFaces: Boolean flag to skip images without faces
* - faceCount: Number of faces detected (0 if no faces)
* - facesLastDetected: Timestamp of last face detection
* - faceDetectionVersion: Version number for cache invalidation
*
* These fields are populated during:
* 1. Initial model training (already detecting faces)
* 2. Utility scans (burst detection, quality analysis)
* 3. Any face detection operation
* 4. Background maintenance scans
*/
@Entity(
tableName = "images",
indices = [
Index(value = ["imageUri"], unique = true),
Index(value = ["sha256"], unique = true),
Index(value = ["capturedAt"]),
Index(value = ["hasFaces"]), // NEW: For fast filtering
Index(value = ["faceCount"]) // NEW: For range queries (singles, couples, groups)
]
)
data class ImageEntity(
@PrimaryKey
val imageId: String,
val imageUri: String,
/**
* Cryptographic hash of image bytes.
* Used for deduplication and re-indexing.
*/
val sha256: String,
/**
* EXIF timestamp (UTC millis).
*/
val capturedAt: Long,
/**
* When this image was indexed into the app.
*/
val ingestedAt: Long,
val width: Int,
val height: Int,
/**
* CAMERA | SCREENSHOT | IMPORTED
*/
val source: String,
// ============================================================================
// FACE DETECTION CACHE - Populated asynchronously
// ============================================================================
/**
* Whether this image contains any faces.
* - true: At least one face detected
* - false: No faces detected
* - null: Not yet scanned (default for newly ingested images)
*
* Use this to skip images without faces during person scanning.
*/
val hasFaces: Boolean? = null,
/**
* Number of faces detected in this image.
* - 0: No faces
* - 1: Solo person (useful for filtering)
* - 2: Couple (useful for filtering)
* - 3+: Group photo (useful for filtering)
* - null: Not yet scanned
*
* Use this for:
* - Finding solo photos of a person
* - Identifying couple photos
* - Filtering out group photos if needed
*/
val faceCount: Int? = null,
/**
* Timestamp when faces were last detected in this image.
* Used for cache invalidation logic.
*
* Invalidate cache if:
* - Image modified date > facesLastDetected
* - faceDetectionVersion < current version
*/
val facesLastDetected: Long? = null,
/**
* Face detection algorithm version.
* Increment this when improving face detection to invalidate old cache.
*
* Current version: 1
* - If detection algorithm improves, increment to 2
* - Query will re-scan images with version < 2
*/
val faceDetectionVersion: Int? = null
) {
companion object {
/**
* Current face detection algorithm version.
* Increment when making significant improvements to face detection.
*/
const val CURRENT_FACE_DETECTION_VERSION = 1
/**
* Check if face detection cache is valid.
* Invalid if:
* - Never scanned (hasFaces == null)
* - Old detection version
* - Image modified after detection (would need file system check)
*/
fun isFaceDetectionCacheValid(image: ImageEntity): Boolean {
return image.hasFaces != null &&
image.faceDetectionVersion == CURRENT_FACE_DETECTION_VERSION
}
}
/**
* Check if this image needs face detection scanning.
*/
fun needsFaceDetection(): Boolean {
return hasFaces == null ||
faceDetectionVersion == null ||
faceDetectionVersion < CURRENT_FACE_DETECTION_VERSION
}
/**
* Check if this image definitely has faces (cached).
*/
fun hasCachedFaces(): Boolean {
return hasFaces == true && !needsFaceDetection()
}
/**
* Check if this image definitely has no faces (cached).
*/
fun hasCachedNoFaces(): Boolean {
return hasFaces == false && !needsFaceDetection()
}
/**
* Get a copy with updated face detection cache.
*/
fun withFaceDetectionCache(
hasFaces: Boolean,
faceCount: Int,
timestamp: Long = System.currentTimeMillis()
): ImageEntity {
return copy(
hasFaces = hasFaces,
faceCount = faceCount,
facesLastDetected = timestamp,
faceDetectionVersion = CURRENT_FACE_DETECTION_VERSION
)
}
}

View File

@@ -0,0 +1,42 @@
package com.placeholder.sherpai2.data.local.entity
import androidx.room.Entity
import androidx.room.ForeignKey
import androidx.room.Index
@Entity(
tableName = "image_events",
primaryKeys = ["imageId", "eventId"],
foreignKeys = [
ForeignKey(
entity = ImageEntity::class,
parentColumns = ["imageId"],
childColumns = ["imageId"],
onDelete = ForeignKey.CASCADE
),
ForeignKey(
entity = EventEntity::class,
parentColumns = ["eventId"],
childColumns = ["eventId"],
onDelete = ForeignKey.CASCADE
)
],
indices = [
Index("eventId")
]
)
data class ImageEventEntity(
val imageId: String,
val eventId: String,
/**
* AUTO | MANUAL
*/
val source: String,
/**
* User override flag.
*/
val override: Boolean
)

View File

@@ -0,0 +1,56 @@
package com.placeholder.sherpai2.data.local.entity
import androidx.room.Entity
import androidx.room.ForeignKey
import androidx.room.Index
/**
* Join table linking images to tags.
*
* This is NOT optional.
* Do not inline tag lists on ImageEntity.
*/
@Entity(
tableName = "image_tags",
primaryKeys = ["imageId", "tagId"],
foreignKeys = [
ForeignKey(
entity = ImageEntity::class,
parentColumns = ["imageId"],
childColumns = ["imageId"],
onDelete = ForeignKey.CASCADE
),
ForeignKey(
entity = TagEntity::class,
parentColumns = ["tagId"],
childColumns = ["tagId"],
onDelete = ForeignKey.CASCADE
)
],
indices = [
Index("tagId"),
Index("imageId")
]
)
data class ImageTagEntity(
val imageId: String,
val tagId: String,
/**
* AUTO | MANUAL
*/
val source: String,
/**
* ML confidence (01).
*/
val confidence: Float,
/**
* PUBLIC | PRIVATE | HIDDEN
*/
val visibility: String,
val createdAt: Long
)

View File

@@ -0,0 +1,143 @@
package com.placeholder.sherpai2.data.local.entity
import androidx.room.ColumnInfo
import androidx.room.Entity
import androidx.room.PrimaryKey
import java.util.UUID
/**
* Tag type constants - MUST be defined BEFORE TagEntity
* to avoid KSP initialization order issues
*/
object TagType {
const val GENERIC = "GENERIC" // User tags
const val SYSTEM = "SYSTEM" // AI/auto tags
const val HIDDEN = "HIDDEN" // Internal
}
/**
* Common system tag values
*/
object SystemTags {
const val HAS_FACES = "has_faces"
const val MULTIPLE_PEOPLE = "multiple_people"
const val LANDSCAPE = "landscape"
const val PORTRAIT = "portrait"
const val LOW_QUALITY = "low_quality"
const val BLURRY = "blurry"
}
/**
* TagEntity - Normalized tag storage
*
* EXPLICIT COLUMN MAPPINGS for KSP compatibility
*/
@Entity(tableName = "tags")
data class TagEntity(
@PrimaryKey
@ColumnInfo(name = "tagId")
val tagId: String,
@ColumnInfo(name = "type")
val type: String,
@ColumnInfo(name = "value")
val value: String,
@ColumnInfo(name = "createdAt")
val createdAt: Long
) {
companion object {
/**
* Create a generic user tag
*/
fun createUserTag(value: String): TagEntity {
return TagEntity(
tagId = UUID.randomUUID().toString(),
type = TagType.GENERIC,
value = value.trim().lowercase(),
createdAt = System.currentTimeMillis()
)
}
/**
* Create a system tag (auto-generated)
*/
fun createSystemTag(value: String): TagEntity {
return TagEntity(
tagId = UUID.randomUUID().toString(),
type = TagType.SYSTEM,
value = value.trim().lowercase(),
createdAt = System.currentTimeMillis()
)
}
/**
* Create hidden tag (internal use)
*/
fun createHiddenTag(value: String): TagEntity {
return TagEntity(
tagId = UUID.randomUUID().toString(),
type = TagType.HIDDEN,
value = value.trim().lowercase(),
createdAt = System.currentTimeMillis()
)
}
}
/**
* Check if this is a user-created tag
*/
fun isUserTag(): Boolean = type == TagType.GENERIC
/**
* Check if this is a system tag
*/
fun isSystemTag(): Boolean = type == TagType.SYSTEM
/**
* Check if this is a hidden tag
*/
fun isHiddenTag(): Boolean = type == TagType.HIDDEN
/**
* Get display value (capitalized for UI)
*/
fun getDisplayValue(): String = value.replaceFirstChar { it.uppercase() }
}
/**
* TagWithUsage - For queries that include usage count
*
* NOT AN ENTITY - just a POJO for query results
* Do NOT add this to @Database entities list!
*/
data class TagWithUsage(
@ColumnInfo(name = "tagId")
val tagId: String,
@ColumnInfo(name = "type")
val type: String,
@ColumnInfo(name = "value")
val value: String,
@ColumnInfo(name = "createdAt")
val createdAt: Long,
@ColumnInfo(name = "usage_count")
val usageCount: Int
) {
/**
* Convert to TagEntity (without usage count)
*/
fun toTagEntity(): TagEntity {
return TagEntity(
tagId = tagId,
type = type,
value = value,
createdAt = createdAt
)
}
}

View File

@@ -0,0 +1,18 @@
package com.placeholder.sherpai2.data.local.model
import androidx.room.ColumnInfo
import androidx.room.Embedded
import com.placeholder.sherpai2.data.local.entity.CollectionEntity
/**
* CollectionWithDetails - Collection with computed preview data
*
* Room maps this directly from query results
*/
data class CollectionWithDetails(
@Embedded
val collection: CollectionEntity,
@ColumnInfo(name = "actualPhotoCount")
val actualPhotoCount: Int
)

View File

@@ -0,0 +1,46 @@
package com.placeholder.sherpai2.data.local.model
import androidx.room.Embedded
import androidx.room.Junction
import androidx.room.Relation
import com.placeholder.sherpai2.data.local.entity.*
/**
* ImageWithEverything - Fully hydrated image with ALL relationships
*
* Room loads this in ONE query using @Transaction!
* NO N+1 problem - all tags and face tags loaded together
*
* Usage:
* - ImageAggregateDao.observeAllImagesWithEverything()
* - Search, Explore, Albums
*/
data class ImageWithEverything(
@Embedded
val image: ImageEntity,
/**
* Tags for this image (via image_tags join table)
* Room automatically joins through ImageTagEntity
*/
@Relation(
parentColumn = "imageId",
entityColumn = "tagId",
associateBy = Junction(
value = ImageTagEntity::class,
parentColumn = "imageId",
entityColumn = "tagId"
)
)
val tags: List<TagEntity>,
/**
* Face tags for this image
* Room automatically loads all PhotoFaceTagEntity for this imageId
*/
@Relation(
parentColumn = "imageId",
entityColumn = "imageId"
)
val faceTags: List<PhotoFaceTagEntity>
)

View File

@@ -0,0 +1,18 @@
package com.placeholder.sherpai2.data.local.model
import androidx.room.Embedded
import androidx.room.Relation
import com.placeholder.sherpai2.data.local.entity.ImageEntity
import com.placeholder.sherpai2.data.local.entity.ImageTagEntity
data class ImageWithTags(
@Embedded
val image: ImageEntity,
@Relation(
parentColumn = "imageId",
entityColumn = "imageId"
)
val tags: List<ImageTagEntity>
)

View File

@@ -1,23 +0,0 @@
package com.placeholder.sherpai2.data.photos
import android.net.Uri
data class Photo(
val id: Long,
val uri: Uri,
val title: String? = null,
val size: Long,
val dateModified: Int
)
data class Album(
val id: String,
val title: String,
val photos: List<Photo>
)
data class AlbumPhoto(
val id: Int,
val imageUrl: String,
val description: String
)

View File

@@ -1,50 +0,0 @@
package com.placeholder.sherpai2.data.repo
import android.content.Context
import android.provider.MediaStore
import android.content.ContentUris
import android.net.Uri
import android.os.Environment
import android.util.Log
import com.placeholder.sherpai2.data.photos.Photo
import kotlinx.coroutines.Dispatchers
import kotlinx.coroutines.withContext
import java.io.File
class PhotoRepository(private val context: Context) {
fun scanExternalStorage(): Result<List<Photo>> {
// Best Practice: Use Environment.getExternalStorageDirectory()
// only as a fallback or starting point for legacy support.
val rootPath = Environment.getExternalStorageDirectory()
return runCatching {
val photos = mutableListOf<Photo>()
if (rootPath.exists() && rootPath.isDirectory) {
// walkTopDown is efficient but can throw AccessDeniedException
rootPath.walkTopDown()
.maxDepth(3) // Performance Best Practice: Don't scan the whole phone
.filter { it.isFile && isImageFile(it.extension) }
.forEach { file ->
photos.add(mapFileToPhoto(file))
}
}
photos
}.onFailure { e ->
Log.e("PhotoRepo", "Failed to scan filesystem", e)
}
}
private fun mapFileToPhoto(file: File): Photo {
return Photo(
id = file.path.hashCode().toLong(),
uri = Uri.fromFile(file),
title = file.name,
size = file.length(),
dateModified = (file.lastModified() / 1000).toInt()
)
}
private fun isImageFile(ext: String) = listOf("jpg", "jpeg", "png").contains(ext.lowercase())
}

View File

@@ -0,0 +1,327 @@
package com.placeholder.sherpai2.data.repository
import com.placeholder.sherpai2.data.local.dao.CollectionDao
import com.placeholder.sherpai2.data.local.dao.ImageAggregateDao
import com.placeholder.sherpai2.data.local.entity.*
import com.placeholder.sherpai2.data.local.model.CollectionWithDetails
import com.placeholder.sherpai2.ui.search.DateRange
import kotlinx.coroutines.flow.Flow
import kotlinx.coroutines.flow.first
import javax.inject.Inject
import javax.inject.Singleton
/**
* CollectionRepository - Business logic for collections
*
* Handles:
* - Creating smart/static collections
* - Evaluating smart collection filters
* - Managing photos in collections
* - Export functionality
*/
@Singleton
class CollectionRepository @Inject constructor(
private val collectionDao: CollectionDao,
private val imageAggregateDao: ImageAggregateDao
) {
// ==========================================
// COLLECTION OPERATIONS
// ==========================================
suspend fun createSmartCollection(
name: String,
description: String?,
includedPeople: Set<String>,
excludedPeople: Set<String>,
includedTags: Set<String>,
excludedTags: Set<String>,
dateRange: DateRange
): String {
// Create collection
val collection = CollectionEntity.createSmart(name, description)
collectionDao.insert(collection)
// Save filters
val filters = mutableListOf<CollectionFilterEntity>()
includedPeople.forEach {
filters.add(
CollectionFilterEntity.create(
collection.collectionId,
"PERSON_INCLUDE",
it
)
)
}
excludedPeople.forEach {
filters.add(
CollectionFilterEntity.create(
collection.collectionId,
"PERSON_EXCLUDE",
it
)
)
}
includedTags.forEach {
filters.add(
CollectionFilterEntity.create(
collection.collectionId,
"TAG_INCLUDE",
it
)
)
}
excludedTags.forEach {
filters.add(
CollectionFilterEntity.create(
collection.collectionId,
"TAG_EXCLUDE",
it
)
)
}
if (dateRange != DateRange.ALL_TIME) {
filters.add(
CollectionFilterEntity.create(
collection.collectionId,
"DATE_RANGE",
dateRange.name
)
)
}
if (filters.isNotEmpty()) {
collectionDao.insertFilters(filters)
}
// Evaluate and populate
evaluateSmartCollection(collection.collectionId)
return collection.collectionId
}
suspend fun createStaticCollection(
name: String,
description: String?,
imageIds: List<String>
): String {
val collection = CollectionEntity.createStatic(name, description, imageIds.size)
collectionDao.insert(collection)
// Add images
val now = System.currentTimeMillis()
val collectionImages = imageIds.mapIndexed { index, imageId ->
CollectionImageEntity(
collectionId = collection.collectionId,
imageId = imageId,
addedAt = now,
sortOrder = index
)
}
collectionDao.addImages(collectionImages)
collectionDao.updatePhotoCount(collection.collectionId, now)
// Set cover image to first image
if (imageIds.isNotEmpty()) {
val firstImage = imageAggregateDao.observeAllImagesWithEverything()
.first()
.find { it.image.imageId == imageIds.first() }
if (firstImage != null) {
collectionDao.updateCoverImage(collection.collectionId, firstImage.image.imageUri, now)
}
}
return collection.collectionId
}
suspend fun deleteCollection(collectionId: String) {
collectionDao.deleteById(collectionId)
}
fun getAllCollections(): Flow<List<CollectionEntity>> {
return collectionDao.getAllCollections()
}
fun getCollection(collectionId: String): Flow<CollectionEntity?> {
return collectionDao.getByIdFlow(collectionId)
}
fun getCollectionWithDetails(collectionId: String): Flow<CollectionWithDetails?> {
return collectionDao.getCollectionWithDetails(collectionId)
}
// ==========================================
// IMAGE MANAGEMENT
// ==========================================
suspend fun addImageToCollection(collectionId: String, imageId: String) {
val now = System.currentTimeMillis()
val count = collectionDao.getPhotoCount(collectionId)
collectionDao.addImage(
CollectionImageEntity(
collectionId = collectionId,
imageId = imageId,
addedAt = now,
sortOrder = count
)
)
collectionDao.updatePhotoCount(collectionId, now)
// Update cover image if this is the first photo
if (count == 0) {
val images = collectionDao.getPreviewImages(collectionId)
if (images.isNotEmpty()) {
collectionDao.updateCoverImage(collectionId, images.first().imageUri, now)
}
}
}
suspend fun removeImageFromCollection(collectionId: String, imageId: String) {
collectionDao.removeImage(collectionId, imageId)
collectionDao.updatePhotoCount(collectionId, System.currentTimeMillis())
}
suspend fun toggleFavorite(imageId: String) {
val favCollection = collectionDao.getFavoriteCollection()
?: run {
// Create favorites collection if it doesn't exist
val fav = CollectionEntity.createFavorite()
collectionDao.insert(fav)
fav
}
val isFavorite = collectionDao.containsImage(favCollection.collectionId, imageId)
if (isFavorite) {
removeImageFromCollection(favCollection.collectionId, imageId)
} else {
addImageToCollection(favCollection.collectionId, imageId)
}
}
suspend fun isFavorite(imageId: String): Boolean {
val favCollection = collectionDao.getFavoriteCollection() ?: return false
return collectionDao.containsImage(favCollection.collectionId, imageId)
}
fun getImagesInCollection(collectionId: String): Flow<List<ImageEntity>> {
return collectionDao.getImagesInCollection(collectionId)
}
// ==========================================
// SMART COLLECTION EVALUATION
// ==========================================
/**
* Re-evaluate a SMART collection's filters and update its images
*/
suspend fun evaluateSmartCollection(collectionId: String) {
val collection = collectionDao.getById(collectionId) ?: return
if (collection.type != "SMART") return
val filters = collectionDao.getFilters(collectionId)
if (filters.isEmpty()) return
// Get all images
val allImages = imageAggregateDao.observeAllImagesWithEverything().first()
// Parse filters
val includedPeople = filters
.filter { it.filterType == "PERSON_INCLUDE" }
.map { it.filterValue }
.toSet()
val excludedPeople = filters
.filter { it.filterType == "PERSON_EXCLUDE" }
.map { it.filterValue }
.toSet()
val includedTags = filters
.filter { it.filterType == "TAG_INCLUDE" }
.map { it.filterValue }
.toSet()
val excludedTags = filters
.filter { it.filterType == "TAG_EXCLUDE" }
.map { it.filterValue }
.toSet()
// Filter images (same logic as SearchViewModel)
val matchingImages = allImages.filter { imageWithEverything ->
// TODO: Apply same Boolean logic as SearchViewModel
// For now, simple tag matching
val imageTags = imageWithEverything.tags.map { it.value }.toSet()
val hasIncludedTags = includedTags.isEmpty() || includedTags.all { it in imageTags }
val hasNoExcludedTags = excludedTags.isEmpty() || excludedTags.none { it in imageTags }
hasIncludedTags && hasNoExcludedTags
}.map { it.image.imageId }
// Update collection
collectionDao.clearAllImages(collectionId)
val now = System.currentTimeMillis()
val collectionImages = matchingImages.mapIndexed { index, imageId ->
CollectionImageEntity(
collectionId = collectionId,
imageId = imageId,
addedAt = now,
sortOrder = index
)
}
if (collectionImages.isNotEmpty()) {
collectionDao.addImages(collectionImages)
// Set cover image to first image
val firstImageId = matchingImages.first()
val firstImage = allImages.find { it.image.imageId == firstImageId }
if (firstImage != null) {
collectionDao.updateCoverImage(collectionId, firstImage.image.imageUri, now)
}
}
collectionDao.updatePhotoCount(collectionId, now)
}
/**
* Re-evaluate all SMART collections
*/
suspend fun evaluateAllSmartCollections() {
val collections = collectionDao.getCollectionsByType("SMART").first()
collections.forEach { collection ->
evaluateSmartCollection(collection.collectionId)
}
}
// ==========================================
// UPDATES
// ==========================================
suspend fun updateCollectionDetails(
collectionId: String,
name: String,
description: String?
) {
collectionDao.updateDetails(collectionId, name, description, System.currentTimeMillis())
}
suspend fun togglePinned(collectionId: String) {
val collection = collectionDao.getById(collectionId) ?: return
collectionDao.updatePinned(
collectionId,
!collection.isPinned,
System.currentTimeMillis()
)
}
}

View File

@@ -0,0 +1,419 @@
package com.placeholder.sherpai2.data.repository
import android.content.Context
import android.graphics.Bitmap
import com.placeholder.sherpai2.data.local.dao.FaceModelDao
import com.placeholder.sherpai2.data.local.dao.ImageDao
import com.placeholder.sherpai2.data.local.dao.PersonDao
import com.placeholder.sherpai2.data.local.dao.PhotoFaceTagDao
import com.placeholder.sherpai2.data.local.entity.*
import com.placeholder.sherpai2.ml.FaceNetModel
import com.placeholder.sherpai2.ui.trainingprep.TrainingSanityChecker
import kotlinx.coroutines.Dispatchers
import kotlinx.coroutines.flow.Flow
import kotlinx.coroutines.flow.map
import kotlinx.coroutines.withContext
import javax.inject.Inject
import javax.inject.Singleton
/**
* FaceRecognitionRepository - Complete face recognition system
*
* USES STRING IDs TO MATCH YOUR SCHEMA:
* - PersonEntity.id: String (UUID)
* - ImageEntity.imageId: String
* - FaceModelEntity.id: String (UUID)
* - PhotoFaceTagEntity.id: String (UUID)
*/
@Singleton
class FaceRecognitionRepository @Inject constructor(
private val context: Context,
private val personDao: PersonDao,
private val imageDao: ImageDao,
private val faceModelDao: FaceModelDao,
private val photoFaceTagDao: PhotoFaceTagDao
) {
private val faceNetModel by lazy { FaceNetModel(context) }
// ======================
// TRAINING OPERATIONS
// ======================
/**
* Create a new person with face model in one operation.
*
* @return PersonId (String UUID)
*/
suspend fun createPersonWithFaceModel(
personName: String,
validImages: List<TrainingSanityChecker.ValidTrainingImage>,
onProgress: (Int, Int) -> Unit = { _, _ -> }
): String = withContext(Dispatchers.IO) {
// Create PersonEntity with UUID
val person = PersonEntity.create(name = personName)
personDao.insert(person)
// Train face model
trainPerson(
personId = person.id,
validImages = validImages,
onProgress = onProgress
)
person.id
}
/**
* Train a face recognition model for an existing person.
*
* @param personId String UUID
* @return Face model ID (String UUID)
*/
suspend fun trainPerson(
personId: String,
validImages: List<TrainingSanityChecker.ValidTrainingImage>,
onProgress: (Int, Int) -> Unit = { _, _ -> }
): String = withContext(Dispatchers.Default) {
val person = personDao.getPersonById(personId)
?: throw IllegalArgumentException("Person with ID $personId not found")
val embeddings = faceNetModel.generateEmbeddingsBatch(
faceBitmaps = validImages.map { it.croppedFaceBitmap },
onProgress = onProgress
)
val personEmbedding = faceNetModel.createPersonModel(embeddings)
val confidences = embeddings.map { embedding ->
faceNetModel.calculateSimilarity(personEmbedding, embedding)
}
val avgConfidence = confidences.average().toFloat()
val faceModel = FaceModelEntity.create(
personId = personId,
embeddingArray = personEmbedding,
trainingImageCount = validImages.size,
averageConfidence = avgConfidence
)
faceModelDao.insertFaceModel(faceModel)
faceModel.id
}
/**
* Retrain face model with additional images.
*/
suspend fun retrainFaceModel(
faceModelId: String,
newFaceImages: List<Bitmap>
) = withContext(Dispatchers.Default) {
val faceModel = faceModelDao.getFaceModelById(faceModelId)
?: throw IllegalArgumentException("Face model $faceModelId not found")
val existingEmbedding = faceModel.getEmbeddingArray()
val newEmbeddings = faceNetModel.generateEmbeddingsBatch(newFaceImages)
val allEmbeddings = listOf(existingEmbedding) + newEmbeddings
val updatedEmbedding = faceNetModel.createPersonModel(allEmbeddings)
val confidences = allEmbeddings.map { embedding ->
faceNetModel.calculateSimilarity(updatedEmbedding, embedding)
}
val avgConfidence = confidences.average().toFloat()
faceModelDao.updateFaceModel(
FaceModelEntity.create(
personId = faceModel.personId,
embeddingArray = updatedEmbedding,
trainingImageCount = faceModel.trainingImageCount + newFaceImages.size,
averageConfidence = avgConfidence
).copy(
id = faceModelId,
createdAt = faceModel.createdAt,
updatedAt = System.currentTimeMillis()
)
)
}
// ======================
// SCANNING / RECOGNITION
// ======================
/**
* Scan an image for faces and tag recognized persons.
*
* ALSO UPDATES FACE DETECTION CACHE for optimization.
*
* @param imageId String (from ImageEntity.imageId)
*/
suspend fun scanImage(
imageId: String,
detectedFaces: List<DetectedFace>,
threshold: Float = FaceNetModel.SIMILARITY_THRESHOLD_HIGH
): List<PhotoFaceTagEntity> = withContext(Dispatchers.Default) {
// OPTIMIZATION: Update face detection cache
// This makes future scans faster by skipping images without faces
withContext(Dispatchers.IO) {
imageDao.updateFaceDetectionCache(
imageId = imageId,
hasFaces = detectedFaces.isNotEmpty(),
faceCount = detectedFaces.size
)
}
val faceModels = faceModelDao.getAllActiveFaceModels()
if (faceModels.isEmpty()) {
return@withContext emptyList()
}
val tags = mutableListOf<PhotoFaceTagEntity>()
for (detectedFace in detectedFaces) {
val faceEmbedding = faceNetModel.generateEmbedding(detectedFace.croppedBitmap)
var bestMatch: Pair<String, Float>? = null
var highestSimilarity = threshold
for (faceModel in faceModels) {
val modelEmbedding = faceModel.getEmbeddingArray()
val similarity = faceNetModel.calculateSimilarity(faceEmbedding, modelEmbedding)
if (similarity > highestSimilarity) {
highestSimilarity = similarity
bestMatch = Pair(faceModel.id, similarity)
}
}
if (bestMatch != null) {
val (faceModelId, confidence) = bestMatch
val tag = PhotoFaceTagEntity.create(
imageId = imageId,
faceModelId = faceModelId,
boundingBox = detectedFace.boundingBox,
confidence = confidence,
faceEmbedding = faceEmbedding
)
tags.add(tag)
faceModelDao.updateLastUsed(faceModelId, System.currentTimeMillis())
}
}
if (tags.isNotEmpty()) {
photoFaceTagDao.insertTags(tags)
}
tags
}
/**
* Recognize a single face bitmap (without saving).
*/
suspend fun recognizeFace(
faceBitmap: Bitmap,
threshold: Float = FaceNetModel.SIMILARITY_THRESHOLD_HIGH
): Pair<String, Float>? = withContext(Dispatchers.Default) {
val faceEmbedding = faceNetModel.generateEmbedding(faceBitmap)
val faceModels = faceModelDao.getAllActiveFaceModels()
val modelEmbeddings = faceModels.map { it.id to it.getEmbeddingArray() }
faceNetModel.findBestMatch(faceEmbedding, modelEmbeddings, threshold)
}
// ======================
// SEARCH / QUERY
// ======================
/**
* Get all images containing a specific person.
*
* @param personId String UUID
*/
suspend fun getImagesForPerson(personId: String): List<ImageEntity> = withContext(Dispatchers.IO) {
val faceModel = faceModelDao.getFaceModelByPersonId(personId)
?: return@withContext emptyList()
val imageIds = photoFaceTagDao.getImageIdsForFaceModel(faceModel.id)
imageDao.getImagesByIds(imageIds)
}
/**
* Get images for person as Flow (reactive).
*/
fun getImagesForPersonFlow(personId: String): Flow<List<ImageEntity>> {
return photoFaceTagDao.getImageIdsForFaceModelFlow(personId)
.map { imageIds ->
imageDao.getImagesByIds(imageIds)
}
}
/**
* Get all persons with face models.
*/
suspend fun getPersonsWithFaceModels(): List<PersonEntity> = withContext(Dispatchers.IO) {
val faceModels = faceModelDao.getAllActiveFaceModels()
val personIds = faceModels.map { it.personId }
personDao.getPersonsByIds(personIds)
}
/**
* Get face detection stats for a person.
*/
suspend fun getPersonFaceStats(personId: String): PersonFaceStats? = withContext(Dispatchers.IO) {
val person = personDao.getPersonById(personId) ?: return@withContext null
val faceModel = faceModelDao.getFaceModelByPersonId(personId) ?: return@withContext null
val imageIds = photoFaceTagDao.getImageIdsForFaceModel(faceModel.id)
val allTags = photoFaceTagDao.getAllTagsForFaceModel(faceModel.id)
val avgConfidence = if (allTags.isNotEmpty()) {
allTags.map { it.confidence }.average().toFloat()
} else {
0f
}
val lastDetected = allTags.maxOfOrNull { it.detectedAt }
PersonFaceStats(
personId = person.id,
personName = person.name,
faceModelId = faceModel.id,
trainingImageCount = faceModel.trainingImageCount,
taggedPhotoCount = imageIds.size,
averageConfidence = avgConfidence,
lastDetectedAt = lastDetected
)
}
/**
* Get face tags for an image.
*/
suspend fun getFaceTagsForImage(imageId: String): List<PhotoFaceTagEntity> {
return photoFaceTagDao.getTagsForImage(imageId)
}
/**
* Get person from a face tag.
*/
suspend fun getPersonForFaceTag(tag: PhotoFaceTagEntity): PersonEntity? = withContext(Dispatchers.IO) {
val faceModel = faceModelDao.getFaceModelById(tag.faceModelId) ?: return@withContext null
personDao.getPersonById(faceModel.personId)
}
/**
* Get face tags with person info for an image.
*/
suspend fun getFaceTagsWithPersons(imageId: String): List<Pair<PhotoFaceTagEntity, PersonEntity>> = withContext(Dispatchers.IO) {
val tags = photoFaceTagDao.getTagsForImage(imageId)
tags.mapNotNull { tag ->
val person = getPersonForFaceTag(tag)
if (person != null) tag to person else null
}
}
// ======================
// VERIFICATION / QUALITY
// ======================
suspend fun verifyFaceTag(tagId: String) {
photoFaceTagDao.markTagAsVerified(
tagId = tagId,
timestamp = System.currentTimeMillis()
)
}
suspend fun getUnverifiedTags(): List<PhotoFaceTagEntity> {
return photoFaceTagDao.getUnverifiedTags()
}
suspend fun getLowConfidenceTags(threshold: Float = 0.7f): List<PhotoFaceTagEntity> {
return photoFaceTagDao.getLowConfidenceTags(threshold)
}
// ======================
// MANAGEMENT
// ======================
suspend fun deleteFaceModel(faceModelId: String) = withContext(Dispatchers.IO) {
photoFaceTagDao.deleteTagsForFaceModel(faceModelId)
faceModelDao.deleteFaceModelById(faceModelId)
}
// Add this method to FaceRecognitionRepository_StringIds.kt
// Replace the existing createPersonWithFaceModel method with this version:
/**
* Create a new person with face model in one operation.
* Now supports full PersonEntity with optional fields.
*
* @param person PersonEntity with name, DOB, relationship, etc.
* @return PersonId (String UUID)
*/
suspend fun createPersonWithFaceModel(
person: PersonEntity,
validImages: List<TrainingSanityChecker.ValidTrainingImage>,
onProgress: (Int, Int) -> Unit = { _, _ -> }
): String = withContext(Dispatchers.IO) {
// Insert person with all fields
personDao.insert(person)
// Train face model
trainPerson(
personId = person.id,
validImages = validImages,
onProgress = onProgress
)
person.id
}
/**
* Get face model by ID
*/
suspend fun getFaceModelById(faceModelId: String): FaceModelEntity? = withContext(Dispatchers.IO) {
faceModelDao.getFaceModelById(faceModelId)
}
suspend fun deleteTagsForImage(imageId: String) {
photoFaceTagDao.deleteTagsForImage(imageId)
}
/**
* Get all image IDs that have been tagged with this face model
* Used for scan optimization (skip already-tagged images)
*/
suspend fun getImageIdsForFaceModel(faceModelId: String): List<String> = withContext(Dispatchers.IO) {
photoFaceTagDao.getImageIdsForFaceModel(faceModelId)
}
fun cleanup() {
faceNetModel.close()
}
}
data class DetectedFace(
val croppedBitmap: Bitmap,
val boundingBox: android.graphics.Rect
)
data class PersonFaceStats(
val personId: String,
val personName: String,
val faceModelId: String,
val trainingImageCount: Int,
val taggedPhotoCount: Int,
val averageConfidence: Float,
val lastDetectedAt: Long?
)

View File

@@ -0,0 +1,380 @@
package com.placeholder.sherpai2.data.service
import android.content.Context
import android.graphics.Bitmap
import android.graphics.Color
import com.placeholder.sherpai2.data.local.dao.ImageTagDao
import com.placeholder.sherpai2.data.local.dao.PersonDao
import com.placeholder.sherpai2.data.local.dao.PhotoFaceTagDao
import com.placeholder.sherpai2.data.local.dao.TagDao
import com.placeholder.sherpai2.data.local.entity.ImageEntity
import com.placeholder.sherpai2.data.local.entity.ImageTagEntity
import com.placeholder.sherpai2.data.local.entity.TagEntity
import com.placeholder.sherpai2.data.repository.DetectedFace
import com.placeholder.sherpai2.util.DiagnosticLogger
import dagger.hilt.android.qualifiers.ApplicationContext
import kotlinx.coroutines.Dispatchers
import kotlinx.coroutines.withContext
import java.util.Calendar
import javax.inject.Inject
import javax.inject.Singleton
import kotlin.math.abs
/**
* AutoTaggingService - Intelligent auto-tagging system
*
* Capabilities:
* - Face-based tags (group_photo, selfie, couple)
* - Scene tags (portrait, landscape, square orientation)
* - Time tags (morning, afternoon, evening, night)
* - Quality tags (high_res, low_res)
* - Relationship tags (family, friend, colleague from PersonEntity)
* - Birthday tags (from PersonEntity DOB)
* - Indoor/Outdoor estimation (basic heuristic)
*/
@Singleton
class AutoTaggingService @Inject constructor(
@ApplicationContext private val context: Context,
private val tagDao: TagDao,
private val imageTagDao: ImageTagDao,
private val photoFaceTagDao: PhotoFaceTagDao,
private val personDao: PersonDao
) {
// ======================
// MAIN AUTO-TAGGING
// ======================
/**
* Auto-tag an image with all applicable system tags
*
* @return Number of tags applied
*/
suspend fun autoTagImage(
imageEntity: ImageEntity,
bitmap: Bitmap,
detectedFaces: List<DetectedFace>
): Int = withContext(Dispatchers.Default) {
val tagsToApply = mutableListOf<String>()
// Face-count based tags
when (detectedFaces.size) {
0 -> { /* No face tags */ }
1 -> {
if (isSelfie(detectedFaces[0], bitmap)) {
tagsToApply.add("selfie")
} else {
tagsToApply.add("single_person")
}
}
2 -> tagsToApply.add("couple")
in 3..5 -> tagsToApply.add("group_photo")
in 6..10 -> {
tagsToApply.add("group_photo")
tagsToApply.add("large_group")
}
else -> {
tagsToApply.add("group_photo")
tagsToApply.add("large_group")
tagsToApply.add("crowd")
}
}
// Orientation tags
val aspectRatio = bitmap.width.toFloat() / bitmap.height.toFloat()
when {
aspectRatio > 1.3f -> tagsToApply.add("landscape")
aspectRatio < 0.77f -> tagsToApply.add("portrait")
else -> tagsToApply.add("square")
}
// Resolution tags
val megapixels = (bitmap.width * bitmap.height) / 1_000_000f
when {
megapixels > 2.0f -> tagsToApply.add("high_res")
megapixels < 0.5f -> tagsToApply.add("low_res")
}
// Time-based tags
val hourOfDay = getHourFromTimestamp(imageEntity.capturedAt)
tagsToApply.add(when (hourOfDay) {
in 5..10 -> "morning"
in 11..16 -> "afternoon"
in 17..20 -> "evening"
else -> "night"
})
// Indoor/Outdoor estimation (only if image is large enough)
if (bitmap.width >= 200 && bitmap.height >= 200) {
val isIndoor = estimateIndoorOutdoor(bitmap)
tagsToApply.add(if (isIndoor) "indoor" else "outdoor")
}
// Apply all tags
var tagsApplied = 0
tagsToApply.forEach { tagName ->
if (applySystemTag(imageEntity.imageId, tagName)) {
tagsApplied++
}
}
DiagnosticLogger.d("AutoTag: Applied $tagsApplied tags to image ${imageEntity.imageId}")
tagsApplied
}
// ======================
// RELATIONSHIP TAGS
// ======================
/**
* Tag all images with a person using their relationship tag
*
* @param personId Person to tag images for
* @return Number of tags applied
*/
suspend fun autoTagRelationshipForPerson(personId: String): Int = withContext(Dispatchers.IO) {
val person = personDao.getPersonById(personId) ?: return@withContext 0
val relationship = person.relationship?.lowercase() ?: return@withContext 0
// Get face model for this person
val faceModels = photoFaceTagDao.getAllTagsForFaceModel(personId)
if (faceModels.isEmpty()) return@withContext 0
val imageIds = faceModels.map { it.imageId }.distinct()
var tagsApplied = 0
imageIds.forEach { imageId ->
if (applySystemTag(imageId, relationship)) {
tagsApplied++
}
}
DiagnosticLogger.i("AutoTag: Applied '$relationship' tag to $tagsApplied images for ${person.name}")
tagsApplied
}
/**
* Tag relationships for ALL persons in database
*/
suspend fun autoTagAllRelationships(): Int = withContext(Dispatchers.IO) {
val persons = personDao.getAllPersons()
var totalTags = 0
persons.forEach { person ->
totalTags += autoTagRelationshipForPerson(person.id)
}
DiagnosticLogger.i("AutoTag: Applied $totalTags relationship tags across ${persons.size} persons")
totalTags
}
// ======================
// BIRTHDAY TAGS
// ======================
/**
* Tag images near a person's birthday
*
* @param personId Person whose birthday to check
* @param daysRange Days before/after birthday to consider (default: 3)
* @return Number of tags applied
*/
suspend fun autoTagBirthdaysForPerson(
personId: String,
daysRange: Int = 3
): Int = withContext(Dispatchers.IO) {
val person = personDao.getPersonById(personId) ?: return@withContext 0
val dateOfBirth = person.dateOfBirth ?: return@withContext 0
// Get all face tags for this person
val faceTags = photoFaceTagDao.getAllTagsForFaceModel(personId)
if (faceTags.isEmpty()) return@withContext 0
var tagsApplied = 0
faceTags.forEach { faceTag ->
// Get the image to check its timestamp
val imageId = faceTag.imageId
// Check if image was captured near birthday
if (isNearBirthday(faceTag.detectedAt, dateOfBirth, daysRange)) {
if (applySystemTag(imageId, "birthday")) {
tagsApplied++
}
}
}
DiagnosticLogger.i("AutoTag: Applied 'birthday' tag to $tagsApplied images for ${person.name}")
tagsApplied
}
/**
* Tag birthdays for ALL persons with DOB
*/
suspend fun autoTagAllBirthdays(daysRange: Int = 3): Int = withContext(Dispatchers.IO) {
val persons = personDao.getAllPersons()
var totalTags = 0
persons.forEach { person ->
if (person.dateOfBirth != null) {
totalTags += autoTagBirthdaysForPerson(person.id, daysRange)
}
}
DiagnosticLogger.i("AutoTag: Applied $totalTags birthday tags")
totalTags
}
// ======================
// HELPER METHODS
// ======================
/**
* Check if an image is a selfie based on face size
*/
private fun isSelfie(face: DetectedFace, bitmap: Bitmap): Boolean {
val boundingBox = face.boundingBox
val faceArea = boundingBox.width() * boundingBox.height()
val imageArea = bitmap.width * bitmap.height
val faceRatio = faceArea.toFloat() / imageArea.toFloat()
// Selfie = face takes up significant portion (>15% of image)
return faceRatio > 0.15f
}
/**
* Get hour of day from timestamp (0-23)
*/
private fun getHourFromTimestamp(timestamp: Long): Int {
return Calendar.getInstance().apply {
timeInMillis = timestamp
}.get(Calendar.HOUR_OF_DAY)
}
/**
* Check if a timestamp is near a birthday
*/
private fun isNearBirthday(
capturedTimestamp: Long,
dobTimestamp: Long,
daysRange: Int
): Boolean {
val capturedCal = Calendar.getInstance().apply { timeInMillis = capturedTimestamp }
val dobCal = Calendar.getInstance().apply { timeInMillis = dobTimestamp }
val capturedMonth = capturedCal.get(Calendar.MONTH)
val capturedDay = capturedCal.get(Calendar.DAY_OF_MONTH)
val dobMonth = dobCal.get(Calendar.MONTH)
val dobDay = dobCal.get(Calendar.DAY_OF_MONTH)
if (capturedMonth == dobMonth) {
return abs(capturedDay - dobDay) <= daysRange
}
// Handle edge case: birthday near end/start of month
// e.g., DOB = Jan 2, captured = Dec 31 (within 3 days)
if (abs(capturedMonth - dobMonth) == 1 || abs(capturedMonth - dobMonth) == 11) {
val daysInCapturedMonth = capturedCal.getActualMaximum(Calendar.DAY_OF_MONTH)
val daysInDobMonth = dobCal.getActualMaximum(Calendar.DAY_OF_MONTH)
if (capturedMonth < dobMonth || (capturedMonth == 11 && dobMonth == 0)) {
// Captured before DOB month
val dayDiff = (daysInCapturedMonth - capturedDay) + dobDay
return dayDiff <= daysRange
} else {
// Captured after DOB month
val dayDiff = (daysInDobMonth - dobDay) + capturedDay
return dayDiff <= daysRange
}
}
return false
}
/**
* Basic indoor/outdoor estimation using brightness and saturation
*
* Heuristic:
* - Outdoor: Higher brightness (>120), Higher saturation (>0.25)
* - Indoor: Lower brightness, Lower saturation
*/
private fun estimateIndoorOutdoor(bitmap: Bitmap): Boolean {
// Sample pixels for analysis (don't process entire image)
val sampleSize = 100
val sampledPixels = mutableListOf<Int>()
val stepX = bitmap.width / sampleSize.coerceAtMost(bitmap.width)
val stepY = bitmap.height / sampleSize.coerceAtMost(bitmap.height)
for (x in 0 until sampleSize.coerceAtMost(bitmap.width)) {
for (y in 0 until sampleSize.coerceAtMost(bitmap.height)) {
val px = (x * stepX).coerceIn(0, bitmap.width - 1)
val py = (y * stepY).coerceIn(0, bitmap.height - 1)
sampledPixels.add(bitmap.getPixel(px, py))
}
}
if (sampledPixels.isEmpty()) return true // Default to indoor if sampling fails
// Calculate average brightness
val avgBrightness = sampledPixels.map { pixel ->
val r = Color.red(pixel)
val g = Color.green(pixel)
val b = Color.blue(pixel)
(r + g + b) / 3.0f
}.average()
// Calculate color saturation
val avgSaturation = sampledPixels.map { pixel ->
val hsv = FloatArray(3)
Color.colorToHSV(pixel, hsv)
hsv[1] // Saturation
}.average()
// Heuristic: Indoor if low brightness OR low saturation
return avgBrightness < 120 || avgSaturation < 0.25
}
/**
* Apply a system tag to an image (helper to avoid duplicates)
*
* @return true if tag was applied, false if already exists
*/
private suspend fun applySystemTag(imageId: String, tagName: String): Boolean {
return withContext(Dispatchers.IO) {
try {
// Get or create tag
val tag = getOrCreateSystemTag(tagName)
// Create image-tag link
val imageTag = ImageTagEntity(
imageId = imageId,
tagId = tag.tagId,
source = "AUTO",
confidence = 1.0f,
visibility = "PUBLIC",
createdAt = System.currentTimeMillis()
)
imageTagDao.upsert(imageTag)
true
} catch (e: Exception) {
DiagnosticLogger.e("Failed to apply tag '$tagName' to image $imageId", e)
false
}
}
}
/**
* Get existing system tag or create new one
*/
private suspend fun getOrCreateSystemTag(tagName: String): TagEntity {
return withContext(Dispatchers.IO) {
tagDao.getByValue(tagName) ?: run {
val newTag = TagEntity.createSystemTag(tagName)
tagDao.insert(newTag)
newTag
}
}
}
}

View File

@@ -0,0 +1,95 @@
package com.placeholder.sherpai2.di
import android.content.Context
import androidx.room.Room
import com.placeholder.sherpai2.data.local.AppDatabase
import com.placeholder.sherpai2.data.local.MIGRATION_7_8
import com.placeholder.sherpai2.data.local.dao.*
import dagger.Module
import dagger.Provides
import dagger.hilt.InstallIn
import dagger.hilt.android.qualifiers.ApplicationContext
import dagger.hilt.components.SingletonComponent
import javax.inject.Singleton
/**
* DatabaseModule - Provides database and ALL DAOs
*
* PHASE 2 UPDATES:
* - Added PersonAgeTagDao
* - Added migration v7→v8 (commented out for development)
*/
@Module
@InstallIn(SingletonComponent::class)
object DatabaseModule {
// ===== DATABASE =====
@Provides
@Singleton
fun provideDatabase(
@ApplicationContext context: Context
): AppDatabase =
Room.databaseBuilder(
context,
AppDatabase::class.java,
"sherpai.db"
)
// DEVELOPMENT MODE: Destructive migration (fresh install on schema change)
.fallbackToDestructiveMigration()
// PRODUCTION MODE: Uncomment this and remove fallbackToDestructiveMigration()
// .addMigrations(MIGRATION_7_8)
.build()
// ===== CORE DAOs =====
@Provides
fun provideImageDao(db: AppDatabase): ImageDao =
db.imageDao()
@Provides
fun provideTagDao(db: AppDatabase): TagDao =
db.tagDao()
@Provides
fun provideEventDao(db: AppDatabase): EventDao =
db.eventDao()
@Provides
fun provideImageEventDao(db: AppDatabase): ImageEventDao =
db.imageEventDao()
@Provides
fun provideImageAggregateDao(db: AppDatabase): ImageAggregateDao =
db.imageAggregateDao()
@Provides
fun provideImageTagDao(db: AppDatabase): ImageTagDao =
db.imageTagDao()
// ===== FACE RECOGNITION DAOs =====
@Provides
fun providePersonDao(db: AppDatabase): PersonDao =
db.personDao()
@Provides
fun provideFaceModelDao(db: AppDatabase): FaceModelDao =
db.faceModelDao()
@Provides
fun providePhotoFaceTagDao(db: AppDatabase): PhotoFaceTagDao =
db.photoFaceTagDao()
@Provides
fun providePersonAgeTagDao(db: AppDatabase): PersonAgeTagDao = // NEW
db.personAgeTagDao()
// ===== COLLECTIONS DAOs =====
@Provides
fun provideCollectionDao(db: AppDatabase): CollectionDao =
db.collectionDao()
}

View File

@@ -0,0 +1,34 @@
package com.placeholder.sherpai2.di
import android.content.Context
import com.placeholder.sherpai2.ml.FaceNetModel
import dagger.Module
import dagger.Provides
import dagger.hilt.InstallIn
import dagger.hilt.android.qualifiers.ApplicationContext
import dagger.hilt.components.SingletonComponent
import javax.inject.Singleton
/**
* MLModule - Provides ML-related dependencies
*
* This module provides FaceNetModel for dependency injection
*/
@Module
@InstallIn(SingletonComponent::class)
object MLModule {
/**
* Provide FaceNetModel singleton
*
* FaceNetModel loads the MobileFaceNet TFLite model and manages
* face embedding generation for recognition.
*/
@Provides
@Singleton
fun provideFaceNetModel(
@ApplicationContext context: Context
): FaceNetModel {
return FaceNetModel(context)
}
}

View File

@@ -0,0 +1,90 @@
package com.placeholder.sherpai2.di
import android.content.Context
import com.placeholder.sherpai2.data.local.dao.FaceModelDao
import com.placeholder.sherpai2.data.local.dao.ImageDao
import com.placeholder.sherpai2.data.local.dao.PersonDao
import com.placeholder.sherpai2.data.local.dao.PhotoFaceTagDao
import com.placeholder.sherpai2.data.repository.FaceRecognitionRepository
import com.placeholder.sherpai2.data.repository.TaggingRepositoryImpl
import com.placeholder.sherpai2.domain.repository.ImageRepository
import com.placeholder.sherpai2.domain.repository.ImageRepositoryImpl
import com.placeholder.sherpai2.domain.repository.TaggingRepository
import dagger.Binds
import dagger.Module
import dagger.Provides
import dagger.hilt.InstallIn
import dagger.hilt.android.qualifiers.ApplicationContext
import dagger.hilt.components.SingletonComponent
import javax.inject.Singleton
/**
* RepositoryModule - Provides repository implementations
*
* UPDATED TO INCLUDE:
* - FaceRecognitionRepository for face recognition operations
*/
@Module
@InstallIn(SingletonComponent::class)
abstract class RepositoryModule {
// ===== EXISTING REPOSITORY BINDINGS =====
@Binds
@Singleton
abstract fun bindImageRepository(
impl: ImageRepositoryImpl
): ImageRepository
@Binds
@Singleton
abstract fun bindTaggingRepository(
impl: TaggingRepositoryImpl
): TaggingRepository
// ===== COMPANION OBJECT FOR PROVIDES =====
companion object {
/**
* Provide FaceRecognitionRepository
*
* Uses @Provides instead of @Binds because it needs Context parameter
* and multiple DAO dependencies
*
* INJECTED DEPENDENCIES:
* - Context: For FaceNetModel initialization
* - PersonDao: Access existing persons
* - ImageDao: Access existing images
* - FaceModelDao: Manage face models
* - PhotoFaceTagDao: Manage photo tags
*
* USAGE IN VIEWMODEL:
* ```
* @HiltViewModel
* class MyViewModel @Inject constructor(
* private val faceRecognitionRepository: FaceRecognitionRepository
* ) : ViewModel() {
* // Use repository methods
* }
* ```
*/
@Provides
@Singleton
fun provideFaceRecognitionRepository(
@ApplicationContext context: Context,
personDao: PersonDao,
imageDao: ImageDao,
faceModelDao: FaceModelDao,
photoFaceTagDao: PhotoFaceTagDao
): FaceRecognitionRepository {
return FaceRecognitionRepository(
context = context,
personDao = personDao,
imageDao = imageDao,
faceModelDao = faceModelDao,
photoFaceTagDao = photoFaceTagDao
)
}
}
}

View File

@@ -1,3 +0,0 @@
package com.placeholder.sherpai2.domain
//fun getAllPhotos(context: Context): List<Photo> {

View File

@@ -1,7 +0,0 @@
package com.placeholder.sherpai2.domain
import android.content.Context
class PhotoDuplicateScanner(private val context: Context) {
}

View File

@@ -0,0 +1,465 @@
package com.placeholder.sherpai2.domain.clustering
import android.content.Context
import android.graphics.Bitmap
import android.graphics.BitmapFactory
import android.net.Uri
import com.google.mlkit.vision.common.InputImage
import com.google.mlkit.vision.face.FaceDetection
import com.google.mlkit.vision.face.FaceDetectorOptions
import com.placeholder.sherpai2.data.local.dao.ImageDao
import com.placeholder.sherpai2.data.local.entity.ImageEntity
import com.placeholder.sherpai2.ml.FaceNetModel
import dagger.hilt.android.qualifiers.ApplicationContext
import kotlinx.coroutines.Dispatchers
import kotlinx.coroutines.async
import kotlinx.coroutines.awaitAll
import kotlinx.coroutines.coroutineScope
import kotlinx.coroutines.sync.Semaphore
import kotlinx.coroutines.withContext
import java.util.concurrent.atomic.AtomicInteger
import javax.inject.Inject
import javax.inject.Singleton
import kotlin.math.sqrt
/**
* FaceClusteringService - Auto-discover people in photo library
*
* STRATEGY:
* 1. Load all images with faces (from cache)
* 2. Detect faces and generate embeddings (parallel)
* 3. DBSCAN clustering on embeddings
* 4. Co-occurrence analysis (faces in same photo)
* 5. Return high-quality clusters (10-100 people typical)
*
* PERFORMANCE:
* - Uses face detection cache (only ~30% of photos)
* - Parallel processing (12 concurrent)
* - Smart sampling (don't need ALL faces for clustering)
* - Result: ~2-5 minutes for 10,000 photo library
*/
@Singleton
class FaceClusteringService @Inject constructor(
@ApplicationContext private val context: Context,
private val imageDao: ImageDao
) {
private val semaphore = Semaphore(12)
/**
* Main clustering entry point
*
* @param maxFacesToCluster Limit for performance (default 2000)
* @param onProgress Progress callback (current, total, message)
*/
suspend fun discoverPeople(
maxFacesToCluster: Int = 2000,
onProgress: (Int, Int, String) -> Unit = { _, _, _ -> }
): ClusteringResult = withContext(Dispatchers.Default) {
onProgress(0, 100, "Loading images with faces...")
// Step 1: Get images with faces (cached, fast!)
val imagesWithFaces = imageDao.getImagesWithFaces()
if (imagesWithFaces.isEmpty()) {
// Check if face cache is populated at all
val totalImages = withContext(Dispatchers.IO) {
imageDao.getImageCount()
}
if (totalImages == 0) {
return@withContext ClusteringResult(
clusters = emptyList(),
totalFacesAnalyzed = 0,
processingTimeMs = 0,
errorMessage = "No photos in library. Please wait for photo ingestion to complete."
)
}
// Images exist but no face cache - need to run PopulateFaceDetectionCacheUseCase first
return@withContext ClusteringResult(
clusters = emptyList(),
totalFacesAnalyzed = 0,
processingTimeMs = 0,
errorMessage = "Face detection cache not ready. Please wait for initial face scan to complete (check MainActivity progress bar)."
)
}
onProgress(10, 100, "Analyzing ${imagesWithFaces.size} photos...")
val startTime = System.currentTimeMillis()
// Step 2: Detect faces and generate embeddings (parallel)
val allFaces = detectFacesInImages(
images = imagesWithFaces.take(1000), // Smart limit: don't need all photos
onProgress = { current, total ->
onProgress(10 + (current * 40 / total), 100, "Detecting faces... $current/$total")
}
)
if (allFaces.isEmpty()) {
return@withContext ClusteringResult(
clusters = emptyList(),
totalFacesAnalyzed = 0,
processingTimeMs = System.currentTimeMillis() - startTime
)
}
onProgress(50, 100, "Clustering ${allFaces.size} faces...")
// Step 3: DBSCAN clustering on embeddings
val rawClusters = performDBSCAN(
faces = allFaces.take(maxFacesToCluster),
epsilon = 0.30f, // BALANCED: Not too strict, not too loose
minPoints = 5 // Minimum 5 photos to form a cluster
)
onProgress(70, 100, "Analyzing relationships...")
// Step 4: Build co-occurrence graph
val coOccurrenceGraph = buildCoOccurrenceGraph(rawClusters)
onProgress(80, 100, "Selecting representative faces...")
// Step 5: Select representative faces for each cluster
val clusters = rawClusters.map { cluster ->
FaceCluster(
clusterId = cluster.clusterId,
faces = cluster.faces,
representativeFaces = selectRepresentativeFaces(cluster.faces, count = 6),
photoCount = cluster.faces.map { it.imageId }.distinct().size,
averageConfidence = cluster.faces.map { it.confidence }.average().toFloat(),
estimatedAge = estimateAge(cluster.faces),
potentialSiblings = findPotentialSiblings(cluster, rawClusters, coOccurrenceGraph)
)
}.sortedByDescending { it.photoCount } // Most frequent first
onProgress(100, 100, "Found ${clusters.size} people!")
ClusteringResult(
clusters = clusters,
totalFacesAnalyzed = allFaces.size,
processingTimeMs = System.currentTimeMillis() - startTime
)
}
/**
* Detect faces in images and generate embeddings (parallel)
*/
private suspend fun detectFacesInImages(
images: List<ImageEntity>,
onProgress: (Int, Int) -> Unit
): List<DetectedFaceWithEmbedding> = coroutineScope {
val detector = com.google.mlkit.vision.face.FaceDetection.getClient(
com.google.mlkit.vision.face.FaceDetectorOptions.Builder()
.setPerformanceMode(com.google.mlkit.vision.face.FaceDetectorOptions.PERFORMANCE_MODE_ACCURATE)
.setMinFaceSize(0.15f)
.build()
)
val faceNetModel = FaceNetModel(context)
val allFaces = mutableListOf<DetectedFaceWithEmbedding>()
val processedCount = java.util.concurrent.atomic.AtomicInteger(0)
try {
val jobs = images.map { image ->
async {
semaphore.acquire()
try {
val faces = detectFacesInImage(image, detector, faceNetModel)
val current = processedCount.incrementAndGet()
if (current % 10 == 0) {
onProgress(current, images.size)
}
faces
} finally {
semaphore.release()
}
}
}
jobs.awaitAll().flatten().also {
allFaces.addAll(it)
}
} finally {
detector.close()
faceNetModel.close()
}
allFaces
}
private suspend fun detectFacesInImage(
image: ImageEntity,
detector: com.google.mlkit.vision.face.FaceDetector,
faceNetModel: FaceNetModel
): List<DetectedFaceWithEmbedding> = withContext(Dispatchers.IO) {
try {
val uri = Uri.parse(image.imageUri)
val bitmap = loadBitmapDownsampled(uri, 512) ?: return@withContext emptyList()
val mlImage = com.google.mlkit.vision.common.InputImage.fromBitmap(bitmap, 0)
val faces = com.google.android.gms.tasks.Tasks.await(detector.process(mlImage))
val result = faces.mapNotNull { face ->
try {
val faceBitmap = Bitmap.createBitmap(
bitmap,
face.boundingBox.left.coerceIn(0, bitmap.width - 1),
face.boundingBox.top.coerceIn(0, bitmap.height - 1),
face.boundingBox.width().coerceAtMost(bitmap.width - face.boundingBox.left),
face.boundingBox.height().coerceAtMost(bitmap.height - face.boundingBox.top)
)
val embedding = faceNetModel.generateEmbedding(faceBitmap)
faceBitmap.recycle()
DetectedFaceWithEmbedding(
imageId = image.imageId,
imageUri = image.imageUri,
capturedAt = image.capturedAt,
embedding = embedding,
boundingBox = face.boundingBox,
confidence = 1.0f // Placeholder
)
} catch (e: Exception) {
null
}
}
bitmap.recycle()
result
} catch (e: Exception) {
emptyList()
}
}
/**
* DBSCAN clustering algorithm
*/
private fun performDBSCAN(
faces: List<DetectedFaceWithEmbedding>,
epsilon: Float,
minPoints: Int
): List<RawCluster> {
val visited = mutableSetOf<Int>()
val clusters = mutableListOf<RawCluster>()
var clusterId = 0
for (i in faces.indices) {
if (i in visited) continue
val neighbors = findNeighbors(i, faces, epsilon)
if (neighbors.size < minPoints) {
visited.add(i)
continue // Noise point
}
// Start new cluster
val cluster = mutableListOf<DetectedFaceWithEmbedding>()
val queue = ArrayDeque(neighbors)
visited.add(i)
cluster.add(faces[i])
while (queue.isNotEmpty()) {
val pointIdx = queue.removeFirst()
if (pointIdx in visited) continue
visited.add(pointIdx)
cluster.add(faces[pointIdx])
val pointNeighbors = findNeighbors(pointIdx, faces, epsilon)
if (pointNeighbors.size >= minPoints) {
queue.addAll(pointNeighbors.filter { it !in visited })
}
}
if (cluster.size >= minPoints) {
clusters.add(RawCluster(clusterId++, cluster))
}
}
return clusters
}
private fun findNeighbors(
pointIdx: Int,
faces: List<DetectedFaceWithEmbedding>,
epsilon: Float
): List<Int> {
val point = faces[pointIdx]
return faces.indices.filter { i ->
i != pointIdx && cosineSimilarity(point.embedding, faces[i].embedding) > (1 - epsilon)
}
}
private fun cosineSimilarity(a: FloatArray, b: FloatArray): Float {
var dotProduct = 0f
var normA = 0f
var normB = 0f
for (i in a.indices) {
dotProduct += a[i] * b[i]
normA += a[i] * a[i]
normB += b[i] * b[i]
}
return dotProduct / (sqrt(normA) * sqrt(normB))
}
/**
* Build co-occurrence graph (faces appearing in same photos)
*/
private fun buildCoOccurrenceGraph(clusters: List<RawCluster>): Map<Int, Map<Int, Int>> {
val graph = mutableMapOf<Int, MutableMap<Int, Int>>()
for (i in clusters.indices) {
graph[i] = mutableMapOf()
val imageIds = clusters[i].faces.map { it.imageId }.toSet()
for (j in clusters.indices) {
if (i == j) continue
val sharedImages = clusters[j].faces.count { it.imageId in imageIds }
if (sharedImages > 0) {
graph[i]!![j] = sharedImages
}
}
}
return graph
}
private fun findPotentialSiblings(
cluster: RawCluster,
allClusters: List<RawCluster>,
coOccurrenceGraph: Map<Int, Map<Int, Int>>
): List<Int> {
val clusterIdx = allClusters.indexOf(cluster)
if (clusterIdx == -1) return emptyList()
val siblings = coOccurrenceGraph[clusterIdx]
?.filter { (_, count) -> count >= 5 } // At least 5 shared photos
?.keys
?.toList()
?: emptyList()
return siblings
}
/**
* Select diverse representative faces for UI display
*/
private fun selectRepresentativeFaces(
faces: List<DetectedFaceWithEmbedding>,
count: Int
): List<DetectedFaceWithEmbedding> {
if (faces.size <= count) return faces
// Time-based sampling: spread across different dates
val sortedByTime = faces.sortedBy { it.capturedAt }
val step = faces.size / count
return (0 until count).map { i ->
sortedByTime[i * step]
}
}
/**
* Estimate if cluster represents a child (based on photo timestamps)
*/
private fun estimateAge(faces: List<DetectedFaceWithEmbedding>): AgeEstimate {
val timestamps = faces.map { it.capturedAt }.sorted()
val span = timestamps.last() - timestamps.first()
val spanYears = span / (365.25 * 24 * 60 * 60 * 1000)
// If face appearance changes over 3+ years, likely a child
return if (spanYears > 3.0) {
AgeEstimate.CHILD
} else {
AgeEstimate.UNKNOWN
}
}
private fun loadBitmapDownsampled(uri: Uri, maxDim: Int): Bitmap? {
return try {
val opts = BitmapFactory.Options().apply { inJustDecodeBounds = true }
context.contentResolver.openInputStream(uri)?.use {
BitmapFactory.decodeStream(it, null, opts)
}
var sample = 1
while (opts.outWidth / sample > maxDim || opts.outHeight / sample > maxDim) {
sample *= 2
}
val finalOpts = BitmapFactory.Options().apply {
inSampleSize = sample
inPreferredConfig = Bitmap.Config.RGB_565
}
context.contentResolver.openInputStream(uri)?.use {
BitmapFactory.decodeStream(it, null, finalOpts)
}
} catch (e: Exception) {
null
}
}
}
// ==================
// DATA CLASSES
// ==================
data class DetectedFaceWithEmbedding(
val imageId: String,
val imageUri: String,
val capturedAt: Long,
val embedding: FloatArray,
val boundingBox: android.graphics.Rect,
val confidence: Float
) {
override fun equals(other: Any?): Boolean {
if (this === other) return true
if (javaClass != other?.javaClass) return false
other as DetectedFaceWithEmbedding
return imageId == other.imageId
}
override fun hashCode(): Int = imageId.hashCode()
}
data class RawCluster(
val clusterId: Int,
val faces: List<DetectedFaceWithEmbedding>
)
data class FaceCluster(
val clusterId: Int,
val faces: List<DetectedFaceWithEmbedding>,
val representativeFaces: List<DetectedFaceWithEmbedding>,
val photoCount: Int,
val averageConfidence: Float,
val estimatedAge: AgeEstimate,
val potentialSiblings: List<Int>
)
data class ClusteringResult(
val clusters: List<FaceCluster>,
val totalFacesAnalyzed: Int,
val processingTimeMs: Long,
val errorMessage: String? = null
)
enum class AgeEstimate {
CHILD, // Appearance changes significantly over time
ADULT, // Stable appearance
UNKNOWN // Not enough data
}

View File

@@ -0,0 +1,87 @@
package com.placeholder.sherpai2.domain.repository
import android.graphics.Bitmap
import android.graphics.BitmapFactory
import android.net.Uri
import com.placeholder.sherpai2.data.local.dao.FaceCacheStats
import com.placeholder.sherpai2.data.local.entity.ImageEntity
import com.placeholder.sherpai2.data.local.model.ImageWithEverything
import kotlinx.coroutines.flow.Flow
/**
* Canonical access point for images.
*
* ViewModels must NEVER talk directly to DAOs.
*/
interface ImageRepository {
/**
* Observe a fully-hydrated image graph.
*
* Used by detail screens.
*/
fun observeImage(imageId: String): Flow<ImageWithEverything>
/**
* Ingest images discovered on device.
*
* This function:
* - deduplicates
* - assigns events automatically
* - BLOCKS until complete (old behavior)
*/
suspend fun ingestImages()
/**
* Ingest images with progress callback (NEW!)
*
* @param onProgress Called with (current, total) for progress updates
*/
suspend fun ingestImagesWithProgress(onProgress: (current: Int, total: Int) -> Unit)
/**
* Get total image count (NEW!)
* Fast query to check if images already loaded
*/
suspend fun getImageCount(): Int
fun getAllImages(): Flow<List<ImageWithEverything>>
fun findImagesByTag(tag: String): Flow<List<ImageWithEverything>>
fun getRecentImages(limit: Int): Flow<List<ImageWithEverything>>
// ==========================================
// FACE DETECTION CACHE - NEW METHODS
// ==========================================
/**
* Update face detection cache for a single image
* Called after detecting faces in an image
*/
suspend fun updateFaceDetectionCache(
imageId: String,
hasFaces: Boolean,
faceCount: Int
)
/**
* Get cache statistics
* Useful for displaying cache coverage in UI
*/
suspend fun getFaceCacheStats(): FaceCacheStats?
/**
* Get images that need face detection
* For background maintenance tasks
*/
suspend fun getImagesNeedingFaceDetection(): List<ImageEntity>
/**
* Load bitmap from URI with optional BitmapFactory.Options
* Used for face detection and other image processing
*/
suspend fun loadBitmap(
uri: Uri,
options: BitmapFactory.Options? = null
): Bitmap?
}

View File

@@ -0,0 +1,237 @@
package com.placeholder.sherpai2.domain.repository
import android.content.ContentUris
import android.content.Context
import android.graphics.Bitmap
import android.graphics.BitmapFactory
import android.net.Uri
import android.provider.MediaStore
import android.util.Log
import com.placeholder.sherpai2.data.local.dao.EventDao
import com.placeholder.sherpai2.data.local.dao.FaceCacheStats
import com.placeholder.sherpai2.data.local.dao.ImageAggregateDao
import com.placeholder.sherpai2.data.local.dao.ImageDao
import com.placeholder.sherpai2.data.local.dao.ImageEventDao
import com.placeholder.sherpai2.data.local.entity.ImageEntity
import com.placeholder.sherpai2.data.local.model.ImageWithEverything
import dagger.hilt.android.qualifiers.ApplicationContext
import kotlinx.coroutines.Dispatchers
import kotlinx.coroutines.flow.Flow
import kotlinx.coroutines.withContext
import kotlinx.coroutines.yield
import java.util.*
import javax.inject.Inject
import javax.inject.Singleton
/**
* ImageRepositoryImpl - SUPER FAST ingestion
*
* OPTIMIZATIONS:
* - Skip SHA256 computation entirely (use URI as unique key)
* - Larger batch sizes (200 instead of 100)
* - Less frequent progress updates
* - No unnecessary string operations
*/
@Singleton
class ImageRepositoryImpl @Inject constructor(
private val imageDao: ImageDao,
private val eventDao: EventDao,
private val imageEventDao: ImageEventDao,
private val aggregateDao: ImageAggregateDao,
@ApplicationContext private val context: Context
) : ImageRepository {
override fun observeImage(imageId: String): Flow<ImageWithEverything> {
return aggregateDao.observeImageWithEverything(imageId)
}
override suspend fun getImageCount(): Int = withContext(Dispatchers.IO) {
return@withContext imageDao.getImageCount()
}
override suspend fun ingestImages(): Unit = withContext(Dispatchers.IO) {
ingestImagesWithProgress { _, _ -> }
}
/**
* OPTIMIZED ingestion - 2-3x faster than before!
*/
override suspend fun ingestImagesWithProgress(
onProgress: (current: Int, total: Int) -> Unit
): Unit = withContext(Dispatchers.IO) {
try {
val projection = arrayOf(
MediaStore.Images.Media._ID,
MediaStore.Images.Media.DATE_TAKEN,
MediaStore.Images.Media.DATE_ADDED,
MediaStore.Images.Media.WIDTH,
MediaStore.Images.Media.HEIGHT,
MediaStore.Images.Media.DATA
)
val sortOrder = "${MediaStore.Images.Media.DATE_ADDED} ASC"
// Count total images
var totalImages = 0
context.contentResolver.query(
MediaStore.Images.Media.EXTERNAL_CONTENT_URI,
arrayOf(MediaStore.Images.Media._ID),
null,
null,
null
)?.use { cursor ->
totalImages = cursor.count
}
if (totalImages == 0) {
Log.i("ImageRepository", "No images found")
return@withContext
}
Log.i("ImageRepository", "Found $totalImages images")
onProgress(0, totalImages)
// LARGER batches for speed
val batchSize = 200
var processed = 0
val ingestTime = System.currentTimeMillis()
context.contentResolver.query(
MediaStore.Images.Media.EXTERNAL_CONTENT_URI,
projection,
null,
null,
sortOrder
)?.use { cursor ->
val idCol = cursor.getColumnIndexOrThrow(MediaStore.Images.Media._ID)
val dateTakenCol = cursor.getColumnIndexOrThrow(MediaStore.Images.Media.DATE_TAKEN)
val dateAddedCol = cursor.getColumnIndexOrThrow(MediaStore.Images.Media.DATE_ADDED)
val widthCol = cursor.getColumnIndexOrThrow(MediaStore.Images.Media.WIDTH)
val heightCol = cursor.getColumnIndexOrThrow(MediaStore.Images.Media.HEIGHT)
val dataCol = cursor.getColumnIndexOrThrow(MediaStore.Images.Media.DATA)
val batch = mutableListOf<ImageEntity>()
while (cursor.moveToNext()) {
val id = cursor.getLong(idCol)
val dateTaken = cursor.getLong(dateTakenCol)
val dateAdded = cursor.getLong(dateAddedCol)
val width = cursor.getInt(widthCol)
val height = cursor.getInt(heightCol)
val filePath = cursor.getString(dataCol) ?: ""
val contentUri = ContentUris.withAppendedId(
MediaStore.Images.Media.EXTERNAL_CONTENT_URI,
id
)
// OPTIMIZATION: Use URI as SHA256 (skip expensive hash computation)
val uriString = contentUri.toString()
val imageEntity = ImageEntity(
imageId = UUID.randomUUID().toString(),
imageUri = uriString,
sha256 = uriString, // Fast! No file I/O
capturedAt = if (dateTaken > 0) dateTaken else dateAdded * 1000,
ingestedAt = ingestTime,
width = width,
height = height,
source = determineSourceFast(filePath)
)
batch.add(imageEntity)
processed++
// Insert batch
if (batch.size >= batchSize) {
imageDao.insertImages(batch)
batch.clear()
// Update progress less frequently (every 200 images)
withContext(Dispatchers.Main) {
onProgress(processed, totalImages)
}
yield()
}
}
// Insert remaining
if (batch.isNotEmpty()) {
imageDao.insertImages(batch)
withContext(Dispatchers.Main) {
onProgress(processed, totalImages)
}
}
}
Log.i("ImageRepository", "Ingestion complete: $processed images")
} catch (e: Exception) {
Log.e("ImageRepository", "Error ingesting images", e)
throw e
}
}
/**
* FAST source determination - no regex, just contains checks
*/
private fun determineSourceFast(filePath: String): String {
return when {
filePath.contains("DCIM", ignoreCase = true) -> "CAMERA"
filePath.contains("Screenshot", ignoreCase = true) -> "SCREENSHOT"
filePath.contains("Download", ignoreCase = true) -> "IMPORTED"
filePath.contains("WhatsApp", ignoreCase = true) -> "IMPORTED"
else -> "CAMERA"
}
}
override fun getAllImages(): Flow<List<ImageWithEverything>> {
return aggregateDao.observeAllImagesWithEverything()
}
override fun findImagesByTag(tag: String): Flow<List<ImageWithEverything>> {
return aggregateDao.observeImagesWithTag(tag)
}
override fun getRecentImages(limit: Int): Flow<List<ImageWithEverything>> {
return imageDao.getRecentImages(limit)
}
// Face detection cache methods
override suspend fun updateFaceDetectionCache(
imageId: String,
hasFaces: Boolean,
faceCount: Int
) = withContext(Dispatchers.IO) {
imageDao.updateFaceDetectionCache(
imageId = imageId,
hasFaces = hasFaces,
faceCount = faceCount,
timestamp = System.currentTimeMillis(),
version = ImageEntity.CURRENT_FACE_DETECTION_VERSION
)
}
override suspend fun getFaceCacheStats(): FaceCacheStats? = withContext(Dispatchers.IO) {
imageDao.getFaceCacheStats()
}
override suspend fun getImagesNeedingFaceDetection(): List<ImageEntity> = withContext(Dispatchers.IO) {
imageDao.getImagesNeedingFaceDetection()
}
override suspend fun loadBitmap(
uri: Uri,
options: BitmapFactory.Options?
): Bitmap? = withContext(Dispatchers.IO) {
try {
context.contentResolver.openInputStream(uri)?.use { stream ->
BitmapFactory.decodeStream(stream, null, options)
}
} catch (e: Exception) {
Log.e("ImageRepository", "Failed to load bitmap from $uri", e)
null
}
}
}

View File

@@ -0,0 +1,124 @@
package com.placeholder.sherpai2.domain.repository
import com.placeholder.sherpai2.data.local.dao.ImageDao
import com.placeholder.sherpai2.data.local.entity.ImageEntity
/**
* Extension functions for ImageRepository to support face detection cache
*
* Add these methods to your ImageRepository interface or implementation
*/
/**
* Update face detection cache for a single image
* Called after detecting faces in an image
*/
suspend fun ImageRepository.updateFaceDetectionCache(
imageId: String,
hasFaces: Boolean,
faceCount: Int
) {
// Assuming you have access to ImageDao in your repository
// Adjust based on your actual repository structure
getImageDao().updateFaceDetectionCache(
imageId = imageId,
hasFaces = hasFaces,
faceCount = faceCount,
timestamp = System.currentTimeMillis(),
version = ImageEntity.CURRENT_FACE_DETECTION_VERSION
)
}
/**
* Get cache statistics
* Useful for displaying cache coverage in UI
*/
suspend fun ImageRepository.getFaceCacheStats() =
getImageDao().getFaceCacheStats()
/**
* Get images that need face detection
* For background maintenance tasks
*/
suspend fun ImageRepository.getImagesNeedingFaceDetection() =
getImageDao().getImagesNeedingFaceDetection()
/**
* Batch populate face detection cache
* For initial cache population or maintenance
*/
suspend fun ImageRepository.populateFaceDetectionCache(
onProgress: (current: Int, total: Int) -> Unit = { _, _ -> }
) {
val imagesToProcess = getImageDao().getImagesNeedingFaceDetection()
val total = imagesToProcess.size
imagesToProcess.forEachIndexed { index, image ->
try {
// Detect faces (implement based on your face detection logic)
val faceCount = detectFaceCount(image.imageUri)
updateFaceDetectionCache(
imageId = image.imageId,
hasFaces = faceCount > 0,
faceCount = faceCount
)
if (index % 10 == 0) {
onProgress(index, total)
}
} catch (e: Exception) {
// Skip errors, continue with next image
}
}
onProgress(total, total)
}
/**
* Helper to get ImageDao from repository
* Adjust based on your actual repository structure
*/
private fun ImageRepository.getImageDao(): ImageDao {
// This assumes your ImageRepository has a reference to ImageDao
// Adjust based on your actual implementation:
// Option 1: If ImageRepository is an interface, add this as a method
// Option 2: If it's a class, access the dao directly
// Option 3: Pass ImageDao as a parameter to these functions
throw NotImplementedError("Implement based on your repository structure")
}
/**
* Helper to detect face count
* Implement based on your face detection logic
*/
private suspend fun ImageRepository.detectFaceCount(imageUri: String): Int {
// Implement your face detection logic here
// This is a placeholder - adjust based on your FaceDetectionHelper
throw NotImplementedError("Implement based on your face detection logic")
}
/**
* ALTERNATIVE: If you prefer to add methods directly to ImageRepository,
* add these to your ImageRepository interface:
*
* interface ImageRepository {
* // ... existing methods
*
* suspend fun updateFaceDetectionCache(
* imageId: String,
* hasFaces: Boolean,
* faceCount: Int
* )
*
* suspend fun getFaceCacheStats(): FaceCacheStats?
*
* suspend fun getImagesNeedingFaceDetection(): List<ImageEntity>
*
* suspend fun populateFaceDetectionCache(
* onProgress: (current: Int, total: Int) -> Unit = { _, _ -> }
* )
* }
*
* Then implement these in your ImageRepositoryImpl class.
*/

View File

@@ -0,0 +1,30 @@
package com.placeholder.sherpai2.domain.repository
import com.placeholder.sherpai2.data.local.entity.TagEntity
import kotlinx.coroutines.flow.Flow
/**
* Handles all tagging operations.
*
* This repository is the ONLY place where:
* - tags are attached
* - visibility rules are applied
*/
interface TaggingRepository {
suspend fun addTagToImage(
imageId: String,
tagValue: String,
source: String,
confidence: Float
)
suspend fun hideTagForImage(
imageId: String,
tagValue: String
)
fun getTagsForImage(imageId: String): Flow<List<TagEntity>>
suspend fun removeTagFromImage(imageId: String, tagId: String)
}

View File

@@ -0,0 +1,97 @@
package com.placeholder.sherpai2.data.repository
import com.placeholder.sherpai2.data.local.dao.ImageTagDao
import com.placeholder.sherpai2.data.local.dao.TagDao
import com.placeholder.sherpai2.data.local.entity.ImageTagEntity
import com.placeholder.sherpai2.data.local.entity.TagEntity
import com.placeholder.sherpai2.domain.repository.TaggingRepository
import kotlinx.coroutines.flow.Flow
import javax.inject.Inject
import javax.inject.Singleton
/**
*
*
* Critical design decisions here
*
* Tag normalization happens once
*
* Visibility rules live here
*
* ML and manual tagging share the same path
*/
@Singleton
class TaggingRepositoryImpl @Inject constructor(
private val tagDao: TagDao,
private val imageTagDao: ImageTagDao
) : TaggingRepository {
override suspend fun addTagToImage(
imageId: String,
tagValue: String,
source: String,
confidence: Float
) {
// Step 1: normalize tag
val normalized = tagValue.trim().lowercase()
// Step 2: ensure tag exists
val tag = tagDao.getByValue(normalized)
?: TagEntity(
tagId = "tag_$normalized",
type = "GENERIC",
value = normalized,
createdAt = System.currentTimeMillis()
).also { tagDao.insert(it) }
// Step 3: attach tag to image
imageTagDao.upsert(
ImageTagEntity(
imageId = imageId,
tagId = tag.tagId,
source = source,
confidence = confidence,
visibility = "PUBLIC",
createdAt = System.currentTimeMillis()
)
)
}
override suspend fun hideTagForImage(
imageId: String,
tagValue: String
) {
val tag = tagDao.getByValue(tagValue) ?: return
imageTagDao.upsert(
ImageTagEntity(
imageId = imageId,
tagId = tag.tagId,
source = "MANUAL",
confidence = 1.0f,
visibility = "HIDDEN",
createdAt = System.currentTimeMillis()
)
)
}
override fun getTagsForImage(imageId: String): Flow<List<TagEntity>> {
// Join imageTagDao -> tagDao to get all PUBLIC tags for this image
return imageTagDao.getTagsForImage(imageId)
}
override suspend fun removeTagFromImage(imageId: String, tagId: String) {
// Mark the tag as hidden instead of deleting, keeping the visibility logic
imageTagDao.upsert(
ImageTagEntity(
imageId = imageId,
tagId = tagId,
source = "MANUAL",
confidence = 1.0f,
visibility = "HIDDEN",
createdAt = System.currentTimeMillis()
)
)
}
}

View File

@@ -0,0 +1,234 @@
package com.placeholder.sherpai2.domain.training
import android.content.Context
import android.graphics.BitmapFactory
import android.net.Uri
import com.placeholder.sherpai2.data.local.dao.FaceModelDao
import com.placeholder.sherpai2.data.local.dao.PersonDao
import com.placeholder.sherpai2.data.local.entity.FaceModelEntity
import com.placeholder.sherpai2.data.local.entity.PersonEntity
import com.placeholder.sherpai2.data.local.entity.TemporalCentroid
import com.placeholder.sherpai2.domain.clustering.FaceCluster
import com.placeholder.sherpai2.ml.FaceNetModel
import dagger.hilt.android.qualifiers.ApplicationContext
import kotlinx.coroutines.Dispatchers
import kotlinx.coroutines.withContext
import javax.inject.Inject
import javax.inject.Singleton
import kotlin.math.abs
/**
* ClusterTrainingService - Train multi-centroid face models from clusters
*
* STRATEGY:
* 1. For children: Create multiple temporal centroids (one per age period)
* 2. For adults: Create single centroid (stable appearance)
* 3. Use K-Means clustering on timestamps to find age groups
* 4. Calculate centroid for each time period
*/
@Singleton
class ClusterTrainingService @Inject constructor(
@ApplicationContext private val context: Context,
private val personDao: PersonDao,
private val faceModelDao: FaceModelDao
) {
private val faceNetModel by lazy { FaceNetModel(context) }
/**
* Train a person from an auto-discovered cluster
*
* @return PersonId on success
*/
suspend fun trainFromCluster(
cluster: FaceCluster,
name: String,
dateOfBirth: Long?,
isChild: Boolean,
siblingClusterIds: List<Int>,
onProgress: (Int, Int, String) -> Unit = { _, _, _ -> }
): String = withContext(Dispatchers.Default) {
onProgress(0, 100, "Creating person...")
// Step 1: Create PersonEntity
val person = PersonEntity.create(
name = name,
dateOfBirth = dateOfBirth,
isChild = isChild,
siblingIds = emptyList(), // Will update after siblings are created
relationship = if (isChild) "Child" else null
)
withContext(Dispatchers.IO) {
personDao.insert(person)
}
onProgress(20, 100, "Analyzing face variations...")
// Step 2: Generate embeddings for all faces in cluster
val facesWithEmbeddings = cluster.faces.mapNotNull { face ->
try {
val bitmap = context.contentResolver.openInputStream(Uri.parse(face.imageUri))?.use {
BitmapFactory.decodeStream(it)
} ?: return@mapNotNull null
// Generate embedding
val embedding = faceNetModel.generateEmbedding(bitmap)
bitmap.recycle()
Triple(face.imageUri, face.capturedAt, embedding)
} catch (e: Exception) {
null
}
}
if (facesWithEmbeddings.isEmpty()) {
throw Exception("Failed to process any faces from cluster")
}
onProgress(50, 100, "Creating face model...")
// Step 3: Create centroids based on whether person is a child
val centroids = if (isChild && dateOfBirth != null) {
createTemporalCentroidsForChild(
facesWithEmbeddings = facesWithEmbeddings,
dateOfBirth = dateOfBirth
)
} else {
createSingleCentroid(facesWithEmbeddings)
}
onProgress(80, 100, "Saving model...")
// Step 4: Calculate average confidence
val avgConfidence = centroids.map { it.avgConfidence }.average().toFloat()
// Step 5: Create FaceModelEntity
val faceModel = FaceModelEntity.createFromCentroids(
personId = person.id,
centroids = centroids,
trainingImageCount = cluster.faces.size,
averageConfidence = avgConfidence
)
withContext(Dispatchers.IO) {
faceModelDao.insertFaceModel(faceModel)
}
onProgress(100, 100, "Complete!")
person.id
}
/**
* Create temporal centroids for a child
* Groups faces by age and creates one centroid per age period
*/
private fun createTemporalCentroidsForChild(
facesWithEmbeddings: List<Triple<String, Long, FloatArray>>,
dateOfBirth: Long
): List<TemporalCentroid> {
// Group faces by age (in years)
val facesByAge = facesWithEmbeddings.groupBy { (_, capturedAt, _) ->
val ageMs = capturedAt - dateOfBirth
val ageYears = (ageMs / (365.25 * 24 * 60 * 60 * 1000)).toInt()
ageYears.coerceIn(0, 18) // Cap at 18 years
}
// Create one centroid per age group
return facesByAge.map { (age, faces) ->
val embeddings = faces.map { it.third }
val avgEmbedding = averageEmbeddings(embeddings)
val avgTimestamp = faces.map { it.second }.average().toLong()
// Calculate confidence (how similar faces are to each other)
val confidences = embeddings.map { emb ->
cosineSimilarity(avgEmbedding, emb)
}
val avgConfidence = confidences.average().toFloat()
TemporalCentroid(
embedding = avgEmbedding.toList(),
effectiveTimestamp = avgTimestamp,
ageAtCapture = age.toFloat(),
photoCount = faces.size,
timeRangeMonths = 12, // 1 year window
avgConfidence = avgConfidence
)
}.sortedBy { it.ageAtCapture }
}
/**
* Create single centroid for an adult (stable appearance)
*/
private fun createSingleCentroid(
facesWithEmbeddings: List<Triple<String, Long, FloatArray>>
): List<TemporalCentroid> {
val embeddings = facesWithEmbeddings.map { it.third }
val avgEmbedding = averageEmbeddings(embeddings)
val avgTimestamp = facesWithEmbeddings.map { it.second }.average().toLong()
val confidences = embeddings.map { emb ->
cosineSimilarity(avgEmbedding, emb)
}
val avgConfidence = confidences.average().toFloat()
return listOf(
TemporalCentroid(
embedding = avgEmbedding.toList(),
effectiveTimestamp = avgTimestamp,
ageAtCapture = null,
photoCount = facesWithEmbeddings.size,
timeRangeMonths = 24, // 2 year window for adults
avgConfidence = avgConfidence
)
)
}
/**
* Average multiple embeddings into one
*/
private fun averageEmbeddings(embeddings: List<FloatArray>): FloatArray {
val size = embeddings.first().size
val avg = FloatArray(size) { 0f }
embeddings.forEach { embedding ->
for (i in embedding.indices) {
avg[i] += embedding[i]
}
}
val count = embeddings.size.toFloat()
for (i in avg.indices) {
avg[i] /= count
}
// Normalize to unit length
val norm = kotlin.math.sqrt(avg.map { it * it }.sum())
return avg.map { it / norm }.toFloatArray()
}
/**
* Calculate cosine similarity between two embeddings
*/
private fun cosineSimilarity(a: FloatArray, b: FloatArray): Float {
var dotProduct = 0f
var normA = 0f
var normB = 0f
for (i in a.indices) {
dotProduct += a[i] * b[i]
normA += a[i] * a[i]
normB += b[i] * b[i]
}
return dotProduct / (kotlin.math.sqrt(normA) * kotlin.math.sqrt(normB))
}
fun cleanup() {
faceNetModel.close()
}
}

View File

@@ -0,0 +1,221 @@
package com.placeholder.sherpai2.domain.usecase
import android.content.Context
import com.placeholder.sherpai2.data.local.dao.ImageDao
import dagger.hilt.android.qualifiers.ApplicationContext
import kotlinx.coroutines.Dispatchers
import kotlinx.coroutines.async
import kotlinx.coroutines.awaitAll
import kotlinx.coroutines.coroutineScope
import kotlinx.coroutines.sync.Semaphore
import kotlinx.coroutines.sync.Mutex
import kotlinx.coroutines.sync.withLock
import kotlinx.coroutines.tasks.await
import kotlinx.coroutines.withContext
import java.util.concurrent.atomic.AtomicInteger
import javax.inject.Inject
import javax.inject.Singleton
/**
* PopulateFaceDetectionCache - HYPER-PARALLEL face scanning
*
* STRATEGY: Use ACCURATE mode BUT with MASSIVE parallelization
* - 50 concurrent detections (not 10!)
* - Semaphore limits to prevent OOM
* - Atomic counters for thread-safe progress
* - Smaller images (768px) for speed without quality loss
*
* RESULT: ~2000-3000 images/minute on modern phones
*/
@Singleton
class PopulateFaceDetectionCacheUseCase @Inject constructor(
@ApplicationContext private val context: Context,
private val imageDao: ImageDao
) {
// Limit concurrent operations to prevent OOM
private val semaphore = Semaphore(50) // 50 concurrent detections!
/**
* HYPER-PARALLEL face detection with ACCURATE mode
*/
suspend fun execute(
onProgress: (Int, Int, String?) -> Unit = { _, _, _ -> }
): Int = withContext(Dispatchers.IO) {
// Create detector with ACCURATE mode but optimized settings
val detector = com.google.mlkit.vision.face.FaceDetection.getClient(
com.google.mlkit.vision.face.FaceDetectorOptions.Builder()
.setPerformanceMode(com.google.mlkit.vision.face.FaceDetectorOptions.PERFORMANCE_MODE_ACCURATE)
.setLandmarkMode(com.google.mlkit.vision.face.FaceDetectorOptions.LANDMARK_MODE_NONE) // Don't need landmarks for cache
.setClassificationMode(com.google.mlkit.vision.face.FaceDetectorOptions.CLASSIFICATION_MODE_NONE) // Don't need classification
.setMinFaceSize(0.1f) // Detect smaller faces
.build()
)
try {
val imagesToScan = imageDao.getImagesNeedingFaceDetection()
if (imagesToScan.isEmpty()) {
return@withContext 0
}
val total = imagesToScan.size
val scanned = AtomicInteger(0)
val pendingUpdates = mutableListOf<CacheUpdate>()
val updatesMutex = kotlinx.coroutines.sync.Mutex()
// Process ALL images in parallel with semaphore control
coroutineScope {
val jobs = imagesToScan.map { image ->
async(Dispatchers.Default) {
semaphore.acquire()
try {
// Load bitmap with medium downsampling (768px = good balance)
val bitmap = loadBitmapOptimized(android.net.Uri.parse(image.imageUri))
if (bitmap == null) {
return@async CacheUpdate(image.imageId, false, 0, image.imageUri)
}
// Detect faces
val inputImage = com.google.mlkit.vision.common.InputImage.fromBitmap(bitmap, 0)
val faces = detector.process(inputImage).await()
bitmap.recycle()
CacheUpdate(
imageId = image.imageId,
hasFaces = faces.isNotEmpty(),
faceCount = faces.size,
imageUri = image.imageUri
)
} catch (e: Exception) {
CacheUpdate(image.imageId, false, 0, image.imageUri)
} finally {
semaphore.release()
// Update progress
val current = scanned.incrementAndGet()
if (current % 50 == 0 || current == total) {
onProgress(current, total, image.imageUri)
}
}
}
}
// Wait for all to complete and collect results
jobs.awaitAll().forEach { update ->
updatesMutex.withLock {
pendingUpdates.add(update)
// Batch write to DB every 100 updates
if (pendingUpdates.size >= 100) {
flushUpdates(pendingUpdates.toList())
pendingUpdates.clear()
}
}
}
// Flush remaining
updatesMutex.withLock {
if (pendingUpdates.isNotEmpty()) {
flushUpdates(pendingUpdates)
}
}
}
scanned.get()
} finally {
detector.close()
}
}
/**
* Optimized bitmap loading with configurable max dimension
*/
private fun loadBitmapOptimized(uri: android.net.Uri, maxDim: Int = 768): android.graphics.Bitmap? {
return try {
// Get dimensions
val options = android.graphics.BitmapFactory.Options().apply {
inJustDecodeBounds = true
}
context.contentResolver.openInputStream(uri)?.use { stream ->
android.graphics.BitmapFactory.decodeStream(stream, null, options)
}
// Calculate sample size
var sampleSize = 1
while (options.outWidth / sampleSize > maxDim ||
options.outHeight / sampleSize > maxDim) {
sampleSize *= 2
}
// Load with sample size
val finalOptions = android.graphics.BitmapFactory.Options().apply {
inSampleSize = sampleSize
inPreferredConfig = android.graphics.Bitmap.Config.ARGB_8888 // Better quality
}
context.contentResolver.openInputStream(uri)?.use { stream ->
android.graphics.BitmapFactory.decodeStream(stream, null, finalOptions)
}
} catch (e: Exception) {
null
}
}
/**
* Batch DB update
*/
private suspend fun flushUpdates(updates: List<CacheUpdate>) = withContext(Dispatchers.IO) {
updates.forEach { update ->
try {
imageDao.updateFaceDetectionCache(
imageId = update.imageId,
hasFaces = update.hasFaces,
faceCount = update.faceCount
)
} catch (e: Exception) {
// Skip failed updates //todo
}
}
}
suspend fun getUncachedImageCount(): Int = withContext(Dispatchers.IO) {
imageDao.getImagesNeedingFaceDetectionCount()
}
suspend fun getCacheStats(): CacheStats = withContext(Dispatchers.IO) {
val stats = imageDao.getFaceCacheStats()
CacheStats(
totalImages = stats?.totalImages ?: 0,
imagesWithFaceCache = stats?.imagesWithFaceCache ?: 0,
imagesWithFaces = stats?.imagesWithFaces ?: 0,
imagesWithoutFaces = stats?.imagesWithoutFaces ?: 0,
needsScanning = stats?.needsScanning ?: 0
)
}
}
private data class CacheUpdate(
val imageId: String,
val hasFaces: Boolean,
val faceCount: Int,
val imageUri: String
)
data class CacheStats(
val totalImages: Int,
val imagesWithFaceCache: Int,
val imagesWithFaces: Int,
val imagesWithoutFaces: Int,
val needsScanning: Int
) {
val cacheProgress: Float
get() = if (totalImages > 0) {
imagesWithFaceCache.toFloat() / totalImages.toFloat()
} else 0f
val isComplete: Boolean
get() = needsScanning == 0
}

View File

@@ -0,0 +1,204 @@
package com.placeholder.sherpai2.ml
import android.content.Context
import android.graphics.Bitmap
import org.tensorflow.lite.Interpreter
import java.io.FileInputStream
import java.nio.ByteBuffer
import java.nio.ByteOrder
import java.nio.MappedByteBuffer
import java.nio.channels.FileChannel
import kotlin.math.sqrt
/**
* FaceNetModel - MobileFaceNet wrapper for face recognition
*
* CLEAN IMPLEMENTATION:
* - All IDs are Strings (matching your schema)
* - Generates 192-dimensional embeddings
* - Cosine similarity for matching
*/
class FaceNetModel(private val context: Context) {
companion object {
private const val MODEL_FILE = "mobilefacenet.tflite"
private const val INPUT_SIZE = 112
private const val EMBEDDING_SIZE = 192
const val SIMILARITY_THRESHOLD_HIGH = 0.7f
const val SIMILARITY_THRESHOLD_MEDIUM = 0.6f
const val SIMILARITY_THRESHOLD_LOW = 0.5f
}
private var interpreter: Interpreter? = null
init {
try {
val model = loadModelFile()
interpreter = Interpreter(model)
} catch (e: Exception) {
throw RuntimeException("Failed to load FaceNet model", e)
}
}
/**
* Load TFLite model from assets
*/
private fun loadModelFile(): MappedByteBuffer {
val fileDescriptor = context.assets.openFd(MODEL_FILE)
val inputStream = FileInputStream(fileDescriptor.fileDescriptor)
val fileChannel = inputStream.channel
val startOffset = fileDescriptor.startOffset
val declaredLength = fileDescriptor.declaredLength
return fileChannel.map(FileChannel.MapMode.READ_ONLY, startOffset, declaredLength)
}
/**
* Generate embedding for a single face
*
* @param faceBitmap Cropped face image (will be resized to 112x112)
* @return 192-dimensional embedding
*/
fun generateEmbedding(faceBitmap: Bitmap): FloatArray {
val resized = Bitmap.createScaledBitmap(faceBitmap, INPUT_SIZE, INPUT_SIZE, true)
val inputBuffer = preprocessImage(resized)
val output = Array(1) { FloatArray(EMBEDDING_SIZE) }
interpreter?.run(inputBuffer, output)
return normalizeEmbedding(output[0])
}
/**
* Generate embeddings for multiple faces (batch processing)
*/
fun generateEmbeddingsBatch(
faceBitmaps: List<Bitmap>,
onProgress: (Int, Int) -> Unit = { _, _ -> }
): List<FloatArray> {
return faceBitmaps.mapIndexed { index, bitmap ->
onProgress(index + 1, faceBitmaps.size)
generateEmbedding(bitmap)
}
}
/**
* Create person model by averaging multiple embeddings
*/
fun createPersonModel(embeddings: List<FloatArray>): FloatArray {
require(embeddings.isNotEmpty()) { "Need at least one embedding" }
val averaged = FloatArray(EMBEDDING_SIZE) { 0f }
embeddings.forEach { embedding ->
for (i in embedding.indices) {
averaged[i] += embedding[i]
}
}
val count = embeddings.size.toFloat()
for (i in averaged.indices) {
averaged[i] /= count
}
return normalizeEmbedding(averaged)
}
/**
* Calculate cosine similarity between two embeddings
* Returns value between -1.0 and 1.0 (higher = more similar)
*/
fun calculateSimilarity(embedding1: FloatArray, embedding2: FloatArray): Float {
require(embedding1.size == EMBEDDING_SIZE && embedding2.size == EMBEDDING_SIZE) {
"Invalid embedding size"
}
var dotProduct = 0f
var norm1 = 0f
var norm2 = 0f
for (i in embedding1.indices) {
dotProduct += embedding1[i] * embedding2[i]
norm1 += embedding1[i] * embedding1[i]
norm2 += embedding2[i] * embedding2[i]
}
return dotProduct / (sqrt(norm1) * sqrt(norm2))
}
/**
* Find best matching face model from a list
*
* @param faceEmbedding Embedding to match
* @param modelEmbeddings List of (modelId: String, embedding: FloatArray)
* @param threshold Minimum similarity threshold
* @return Pair of (modelId: String, confidence: Float) or null
*/
fun findBestMatch(
faceEmbedding: FloatArray,
modelEmbeddings: List<Pair<String, FloatArray>>,
threshold: Float = SIMILARITY_THRESHOLD_HIGH
): Pair<String, Float>? {
var bestMatch: Pair<String, Float>? = null
var highestSimilarity = threshold
for ((modelId, modelEmbedding) in modelEmbeddings) {
val similarity = calculateSimilarity(faceEmbedding, modelEmbedding)
if (similarity > highestSimilarity) {
highestSimilarity = similarity
bestMatch = Pair(modelId, similarity)
}
}
return bestMatch
}
/**
* Preprocess image for model input
*/
private fun preprocessImage(bitmap: Bitmap): ByteBuffer {
val buffer = ByteBuffer.allocateDirect(4 * INPUT_SIZE * INPUT_SIZE * 3)
buffer.order(ByteOrder.nativeOrder())
val pixels = IntArray(INPUT_SIZE * INPUT_SIZE)
bitmap.getPixels(pixels, 0, INPUT_SIZE, 0, 0, INPUT_SIZE, INPUT_SIZE)
for (pixel in pixels) {
val r = ((pixel shr 16) and 0xFF) / 255.0f
val g = ((pixel shr 8) and 0xFF) / 255.0f
val b = (pixel and 0xFF) / 255.0f
buffer.putFloat((r - 0.5f) / 0.5f)
buffer.putFloat((g - 0.5f) / 0.5f)
buffer.putFloat((b - 0.5f) / 0.5f)
}
return buffer
}
/**
* Normalize embedding to unit length
*/
private fun normalizeEmbedding(embedding: FloatArray): FloatArray {
var norm = 0f
for (value in embedding) {
norm += value * value
}
norm = sqrt(norm)
return if (norm > 0) {
FloatArray(embedding.size) { i -> embedding[i] / norm }
} else {
embedding
}
}
/**
* Clean up resources
*/
fun close() {
interpreter?.close()
interpreter = null
}
}

View File

@@ -0,0 +1,127 @@
package com.placeholder.sherpai2.ml
/**
* ThresholdStrategy - Smart threshold selection for face recognition
*
* Considers:
* - Training image count
* - Image quality
* - Detection context (group photo, selfie, etc.)
*/
object ThresholdStrategy {
/**
* Get optimal threshold for face recognition
*
* @param trainingCount Number of images used to train the model
* @param imageQuality Quality assessment of the image being scanned
* @param detectionContext Context of the detection (group, selfie, etc.)
* @return Similarity threshold (0.0 - 1.0)
*/
fun getOptimalThreshold(
trainingCount: Int,
imageQuality: ImageQuality = ImageQuality.UNKNOWN,
detectionContext: DetectionContext = DetectionContext.GENERAL
): Float {
// Base threshold from training count
val baseThreshold = when {
trainingCount >= 40 -> 0.68f // High confidence - strict
trainingCount >= 30 -> 0.62f // Good confidence - moderate-strict
trainingCount >= 20 -> 0.56f // Moderate confidence
trainingCount >= 15 -> 0.50f // Acceptable confidence - lenient
else -> 0.48f // Sparse training - very lenient
}
// Adjust based on image quality
val qualityAdjustment = when (imageQuality) {
ImageQuality.HIGH -> -0.02f // Can be stricter with good quality
ImageQuality.MEDIUM -> 0f // No change
ImageQuality.LOW -> +0.03f // Be more lenient with poor quality
ImageQuality.UNKNOWN -> 0f // No change
}
// Adjust based on detection context
val contextAdjustment = when (detectionContext) {
DetectionContext.GROUP_PHOTO -> +0.02f // More lenient in groups (faces smaller)
DetectionContext.SELFIE -> -0.03f // Stricter for close-ups (more detail)
DetectionContext.PROFILE -> +0.02f // More lenient for side profiles
DetectionContext.DISTANT -> +0.03f // More lenient for far away faces
DetectionContext.GENERAL -> 0f // No change
}
// Combine adjustments and clamp to safe range
return (baseThreshold + qualityAdjustment + contextAdjustment).coerceIn(0.40f, 0.75f)
}
/**
* Get threshold for liberal matching (e.g., during testing)
*/
fun getLiberalThreshold(trainingCount: Int): Float {
return when {
trainingCount >= 30 -> 0.52f
trainingCount >= 20 -> 0.48f
else -> 0.45f
}.coerceIn(0.40f, 0.65f)
}
/**
* Get threshold for conservative matching (minimize false positives)
*/
fun getConservativeThreshold(trainingCount: Int): Float {
return when {
trainingCount >= 40 -> 0.72f
trainingCount >= 30 -> 0.68f
trainingCount >= 20 -> 0.62f
else -> 0.58f
}.coerceIn(0.55f, 0.75f)
}
/**
* Estimate image quality from bitmap properties
*/
fun estimateImageQuality(width: Int, height: Int, fileSize: Long = 0): ImageQuality {
val megapixels = (width * height) / 1_000_000f
return when {
megapixels > 4.0f -> ImageQuality.HIGH
megapixels > 1.0f -> ImageQuality.MEDIUM
else -> ImageQuality.LOW
}
}
/**
* Estimate detection context from face count and face size
*/
fun estimateDetectionContext(
faceCount: Int,
faceAreaRatio: Float = 0f
): DetectionContext {
return when {
faceCount == 1 && faceAreaRatio > 0.15f -> DetectionContext.SELFIE
faceCount == 1 && faceAreaRatio < 0.05f -> DetectionContext.DISTANT
faceCount >= 3 -> DetectionContext.GROUP_PHOTO
else -> DetectionContext.GENERAL
}
}
}
/**
* Image quality assessment
*/
enum class ImageQuality {
HIGH, // > 4MP, good lighting
MEDIUM, // 1-4MP
LOW, // < 1MP, poor quality
UNKNOWN // Cannot determine
}
/**
* Detection context
*/
enum class DetectionContext {
GROUP_PHOTO, // Multiple faces (3+)
SELFIE, // Single face, close-up
PROFILE, // Side view
DISTANT, // Face is small in frame
GENERAL // Default
}

View File

@@ -0,0 +1,320 @@
package com.placeholder.sherpai2.ui.album
import androidx.lifecycle.SavedStateHandle
import androidx.lifecycle.ViewModel
import androidx.lifecycle.viewModelScope
import com.placeholder.sherpai2.data.local.dao.ImageDao
import com.placeholder.sherpai2.data.local.dao.ImageTagDao
import com.placeholder.sherpai2.data.local.dao.PersonDao
import com.placeholder.sherpai2.data.local.dao.TagDao
import com.placeholder.sherpai2.data.local.entity.ImageEntity
import com.placeholder.sherpai2.data.local.entity.PersonEntity
import com.placeholder.sherpai2.data.local.entity.PhotoFaceTagEntity
import com.placeholder.sherpai2.data.repository.FaceRecognitionRepository
import com.placeholder.sherpai2.ui.search.DateRange
import dagger.hilt.android.lifecycle.HiltViewModel
import kotlinx.coroutines.flow.*
import kotlinx.coroutines.launch
import java.util.Calendar
import javax.inject.Inject
/**
* AlbumViewModel - Display photos from a specific album (tag, person, or time range)
*
* Features:
* - Search within album
* - Date filtering
* - Album stats
* - Export functionality
*/
@HiltViewModel
class AlbumViewModel @Inject constructor(
savedStateHandle: SavedStateHandle,
private val tagDao: TagDao,
private val imageTagDao: ImageTagDao,
private val imageDao: ImageDao,
private val personDao: PersonDao,
private val faceRecognitionRepository: FaceRecognitionRepository
) : ViewModel() {
// Album parameters from navigation
private val albumType: String = savedStateHandle["albumType"] ?: "tag"
private val albumId: String = savedStateHandle["albumId"] ?: ""
// UI state
private val _uiState = MutableStateFlow<AlbumUiState>(AlbumUiState.Loading)
val uiState: StateFlow<AlbumUiState> = _uiState.asStateFlow()
// Search query within album
private val _searchQuery = MutableStateFlow("")
val searchQuery: StateFlow<String> = _searchQuery.asStateFlow()
// Date range filter
private val _dateRange = MutableStateFlow(DateRange.ALL_TIME)
val dateRange: StateFlow<DateRange> = _dateRange.asStateFlow()
init {
loadAlbumData()
}
/**
* Load album data based on type
*/
private fun loadAlbumData() {
viewModelScope.launch {
try {
_uiState.value = AlbumUiState.Loading
when (albumType) {
"tag" -> loadTagAlbum()
"person" -> loadPersonAlbum()
"time" -> loadTimeAlbum()
else -> _uiState.value = AlbumUiState.Error("Unknown album type")
}
} catch (e: Exception) {
_uiState.value = AlbumUiState.Error(e.message ?: "Failed to load album")
}
}
}
private suspend fun loadTagAlbum() {
val tag = tagDao.getByValue(albumId)
if (tag == null) {
_uiState.value = AlbumUiState.Error("Tag not found")
return
}
combine(
_searchQuery,
_dateRange
) { query: String, dateRange: DateRange ->
Pair(query, dateRange)
}.collectLatest { (query, dateRange) ->
val imageIds = imageTagDao.findImagesByTag(tag.tagId, 0.5f)
val images = imageDao.getImagesByIds(imageIds)
val filteredImages = images
.filter { isInDateRange(it.capturedAt, dateRange) }
.filter {
query.isBlank() || containsQuery(it, query)
}
val imagesWithFaces = filteredImages.map { image ->
val tagsWithPersons = faceRecognitionRepository.getFaceTagsWithPersons(image.imageId)
AlbumPhoto(
image = image,
faceTags = tagsWithPersons.map { it.first },
persons = tagsWithPersons.map { it.second }
)
}
val uniquePersons = imagesWithFaces
.flatMap { it.persons }
.distinctBy { it.id }
_uiState.value = AlbumUiState.Success(
albumName = tag.value.replace("_", " ").replaceFirstChar { it.uppercase() },
albumType = "Tag",
photos = imagesWithFaces,
personCount = uniquePersons.size,
totalFaces = imagesWithFaces.sumOf { it.faceTags.size }
)
}
}
private suspend fun loadPersonAlbum() {
val person = personDao.getPersonById(albumId)
if (person == null) {
_uiState.value = AlbumUiState.Error("Person not found")
return
}
combine(
_searchQuery,
_dateRange
) { query: String, dateRange: DateRange ->
Pair(query, dateRange)
}.collectLatest { (query, dateRange) ->
val images = faceRecognitionRepository.getImagesForPerson(albumId)
val filteredImages = images
.filter { isInDateRange(it.capturedAt, dateRange) }
.filter {
query.isBlank() || containsQuery(it, query)
}
val imagesWithFaces = filteredImages.map { image ->
val tagsWithPersons = faceRecognitionRepository.getFaceTagsWithPersons(image.imageId)
AlbumPhoto(
image = image,
faceTags = tagsWithPersons.map { it.first },
persons = tagsWithPersons.map { it.second }
)
}
_uiState.value = AlbumUiState.Success(
albumName = person.name,
albumType = "Person",
photos = imagesWithFaces,
personCount = 1,
totalFaces = imagesWithFaces.sumOf { it.faceTags.size }
)
}
}
private suspend fun loadTimeAlbum() {
// Time-based albums (Today, This Week, etc)
val (startTime, endTime, albumName) = when (albumId) {
"today" -> Triple(getStartOfDay(), System.currentTimeMillis(), "Today")
"week" -> Triple(getStartOfWeek(), System.currentTimeMillis(), "This Week")
"month" -> Triple(getStartOfMonth(), System.currentTimeMillis(), "This Month")
"year" -> Triple(getStartOfYear(), System.currentTimeMillis(), "This Year")
else -> {
_uiState.value = AlbumUiState.Error("Unknown time range")
return
}
}
combine(
_searchQuery,
_dateRange
) { query: String, _: DateRange ->
query
}.collectLatest { query ->
val images = imageDao.getImagesInRange(startTime, endTime)
val filteredImages = images.filter {
query.isBlank() || containsQuery(it, query)
}
val imagesWithFaces = filteredImages.map { image ->
val tagsWithPersons = faceRecognitionRepository.getFaceTagsWithPersons(image.imageId)
AlbumPhoto(
image = image,
faceTags = tagsWithPersons.map { it.first },
persons = tagsWithPersons.map { it.second }
)
}
val uniquePersons = imagesWithFaces
.flatMap { it.persons }
.distinctBy { it.id }
_uiState.value = AlbumUiState.Success(
albumName = albumName,
albumType = "Time",
photos = imagesWithFaces,
personCount = uniquePersons.size,
totalFaces = imagesWithFaces.sumOf { it.faceTags.size }
)
}
}
fun setSearchQuery(query: String) {
_searchQuery.value = query
}
fun setDateRange(range: DateRange) {
_dateRange.value = range
}
private fun isInDateRange(timestamp: Long, range: DateRange): Boolean {
return when (range) {
DateRange.ALL_TIME -> true
DateRange.TODAY -> isToday(timestamp)
DateRange.THIS_WEEK -> isThisWeek(timestamp)
DateRange.THIS_MONTH -> isThisMonth(timestamp)
DateRange.THIS_YEAR -> isThisYear(timestamp)
}
}
private fun containsQuery(image: ImageEntity, query: String): Boolean {
// Could expand to search by person names, tags, etc.
return true
}
private fun isToday(timestamp: Long): Boolean {
val today = Calendar.getInstance()
val date = Calendar.getInstance().apply { timeInMillis = timestamp }
return today.get(Calendar.YEAR) == date.get(Calendar.YEAR) &&
today.get(Calendar.DAY_OF_YEAR) == date.get(Calendar.DAY_OF_YEAR)
}
private fun isThisWeek(timestamp: Long): Boolean {
val today = Calendar.getInstance()
val date = Calendar.getInstance().apply { timeInMillis = timestamp }
return today.get(Calendar.YEAR) == date.get(Calendar.YEAR) &&
today.get(Calendar.WEEK_OF_YEAR) == date.get(Calendar.WEEK_OF_YEAR)
}
private fun isThisMonth(timestamp: Long): Boolean {
val today = Calendar.getInstance()
val date = Calendar.getInstance().apply { timeInMillis = timestamp }
return today.get(Calendar.YEAR) == date.get(Calendar.YEAR) &&
today.get(Calendar.MONTH) == date.get(Calendar.MONTH)
}
private fun isThisYear(timestamp: Long): Boolean {
val today = Calendar.getInstance()
val date = Calendar.getInstance().apply { timeInMillis = timestamp }
return today.get(Calendar.YEAR) == date.get(Calendar.YEAR)
}
private fun getStartOfDay(): Long {
return Calendar.getInstance().apply {
set(Calendar.HOUR_OF_DAY, 0)
set(Calendar.MINUTE, 0)
set(Calendar.SECOND, 0)
set(Calendar.MILLISECOND, 0)
}.timeInMillis
}
private fun getStartOfWeek(): Long {
return Calendar.getInstance().apply {
set(Calendar.DAY_OF_WEEK, firstDayOfWeek)
set(Calendar.HOUR_OF_DAY, 0)
set(Calendar.MINUTE, 0)
set(Calendar.SECOND, 0)
set(Calendar.MILLISECOND, 0)
}.timeInMillis
}
private fun getStartOfMonth(): Long {
return Calendar.getInstance().apply {
set(Calendar.DAY_OF_MONTH, 1)
set(Calendar.HOUR_OF_DAY, 0)
set(Calendar.MINUTE, 0)
set(Calendar.SECOND, 0)
set(Calendar.MILLISECOND, 0)
}.timeInMillis
}
private fun getStartOfYear(): Long {
return Calendar.getInstance().apply {
set(Calendar.DAY_OF_YEAR, 1)
set(Calendar.HOUR_OF_DAY, 0)
set(Calendar.MINUTE, 0)
set(Calendar.SECOND, 0)
set(Calendar.MILLISECOND, 0)
}.timeInMillis
}
}
sealed class AlbumUiState {
object Loading : AlbumUiState()
data class Success(
val albumName: String,
val albumType: String,
val photos: List<AlbumPhoto>,
val personCount: Int,
val totalFaces: Int
) : AlbumUiState()
data class Error(val message: String) : AlbumUiState()
}
data class AlbumPhoto(
val image: ImageEntity,
val faceTags: List<PhotoFaceTagEntity>,
val persons: List<PersonEntity>
)

View File

@@ -0,0 +1,481 @@
package com.placeholder.sherpai2.ui.album
import androidx.compose.foundation.clickable
import androidx.compose.foundation.layout.*
import androidx.compose.foundation.lazy.LazyRow
import androidx.compose.foundation.lazy.grid.*
import androidx.compose.foundation.shape.RoundedCornerShape
import androidx.compose.material.icons.Icons
import androidx.compose.material.icons.filled.*
import androidx.compose.material3.*
import androidx.compose.runtime.*
import androidx.compose.ui.Alignment
import androidx.compose.ui.Modifier
import androidx.compose.ui.text.font.FontWeight
import androidx.compose.ui.text.style.TextOverflow
import androidx.compose.ui.unit.dp
import androidx.hilt.navigation.compose.hiltViewModel
import androidx.lifecycle.compose.collectAsStateWithLifecycle
import coil.compose.AsyncImage
import com.placeholder.sherpai2.ui.search.DateRange
/**
* AlbumViewScreen - CLEAN VERSION with Export
*
* REMOVED:
* - DisplayMode toggle
* - Verbose person tags
*
* ADDED:
* - Export menu (Folder, Zip, Collage)
* - Clean simple layout
*/
@OptIn(ExperimentalMaterial3Api::class)
@Composable
fun AlbumViewScreen(
onBack: () -> Unit,
onImageClick: (String) -> Unit,
viewModel: AlbumViewModel = hiltViewModel()
) {
val uiState by viewModel.uiState.collectAsStateWithLifecycle()
val searchQuery by viewModel.searchQuery.collectAsStateWithLifecycle()
val dateRange by viewModel.dateRange.collectAsStateWithLifecycle()
var showExportMenu by remember { mutableStateOf(false) }
Scaffold(
topBar = {
TopAppBar(
title = {
Column {
when (val state = uiState) {
is AlbumUiState.Success -> {
Text(
text = state.albumName,
style = MaterialTheme.typography.titleLarge,
fontWeight = FontWeight.Bold
)
Text(
text = "${state.photos.size} photos",
style = MaterialTheme.typography.bodySmall,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
}
else -> {
Text("Album")
}
}
}
},
navigationIcon = {
IconButton(onClick = onBack) {
Icon(Icons.Default.ArrowBack, "Back")
}
},
actions = {
// Export button
IconButton(onClick = { showExportMenu = true }) {
Icon(Icons.Default.FileDownload, "Export")
}
}
)
}
) { paddingValues ->
when (val state = uiState) {
is AlbumUiState.Loading -> {
Box(
modifier = Modifier
.fillMaxSize()
.padding(paddingValues),
contentAlignment = Alignment.Center
) {
CircularProgressIndicator()
}
}
is AlbumUiState.Error -> {
Box(
modifier = Modifier
.fillMaxSize()
.padding(paddingValues),
contentAlignment = Alignment.Center
) {
Column(
horizontalAlignment = Alignment.CenterHorizontally,
verticalArrangement = Arrangement.spacedBy(8.dp)
) {
Icon(
Icons.Default.Error,
contentDescription = null,
modifier = Modifier.size(48.dp),
tint = MaterialTheme.colorScheme.error
)
Text(state.message)
Button(onClick = onBack) {
Text("Go Back")
}
}
}
}
is AlbumUiState.Success -> {
AlbumContent(
state = state,
searchQuery = searchQuery,
dateRange = dateRange,
onSearchChange = { viewModel.setSearchQuery(it) },
onDateRangeChange = { viewModel.setDateRange(it) },
onImageClick = onImageClick,
modifier = Modifier.padding(paddingValues)
)
}
}
}
// Export menu dialog
if (showExportMenu) {
ExportDialog(
albumName = when (val state = uiState) {
is AlbumUiState.Success -> state.albumName
else -> "Album"
},
photoCount = when (val state = uiState) {
is AlbumUiState.Success -> state.photos.size
else -> 0
},
onDismiss = { showExportMenu = false },
onExportToFolder = {
// TODO: Implement folder export
showExportMenu = false
},
onExportToZip = {
// TODO: Implement zip export
showExportMenu = false
},
onExportToCollage = {
// TODO: Implement collage export
showExportMenu = false
}
)
}
}
@Composable
private fun AlbumContent(
state: AlbumUiState.Success,
searchQuery: String,
dateRange: DateRange,
onSearchChange: (String) -> Unit,
onDateRangeChange: (DateRange) -> Unit,
onImageClick: (String) -> Unit,
modifier: Modifier = Modifier
) {
Column(
modifier = modifier.fillMaxSize()
) {
// Stats card
Card(
modifier = Modifier
.fillMaxWidth()
.padding(16.dp),
colors = CardDefaults.cardColors(
containerColor = MaterialTheme.colorScheme.primaryContainer.copy(alpha = 0.3f)
)
) {
Row(
modifier = Modifier
.fillMaxWidth()
.padding(16.dp),
horizontalArrangement = Arrangement.SpaceAround
) {
StatItem(Icons.Default.Photo, "Photos", state.photos.size.toString())
if (state.totalFaces > 0) {
StatItem(Icons.Default.Face, "Faces", state.totalFaces.toString())
}
if (state.personCount > 0) {
StatItem(Icons.Default.People, "People", state.personCount.toString())
}
}
}
// Search bar
OutlinedTextField(
value = searchQuery,
onValueChange = onSearchChange,
placeholder = { Text("Search in album...") },
leadingIcon = { Icon(Icons.Default.Search, null) },
trailingIcon = {
if (searchQuery.isNotEmpty()) {
IconButton(onClick = { onSearchChange("") }) {
Icon(Icons.Default.Clear, "Clear")
}
}
},
modifier = Modifier
.fillMaxWidth()
.padding(horizontal = 16.dp),
singleLine = true,
shape = RoundedCornerShape(16.dp)
)
Spacer(Modifier.height(8.dp))
// Date filters
LazyRow(
modifier = Modifier
.fillMaxWidth()
.padding(horizontal = 16.dp),
horizontalArrangement = Arrangement.spacedBy(8.dp)
) {
items(DateRange.entries.size) { index ->
val range = DateRange.entries[index]
val isActive = dateRange == range
FilterChip(
selected = isActive,
onClick = { onDateRangeChange(range) },
label = { Text(range.displayName) }
)
}
}
Spacer(Modifier.height(8.dp))
// Photo grid
if (state.photos.isEmpty()) {
Box(
modifier = Modifier.fillMaxSize(),
contentAlignment = Alignment.Center
) {
Text(
text = "No photos in this album",
style = MaterialTheme.typography.bodyLarge,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
}
} else {
LazyVerticalGrid(
columns = GridCells.Adaptive(120.dp),
contentPadding = PaddingValues(12.dp),
verticalArrangement = Arrangement.spacedBy(12.dp),
horizontalArrangement = Arrangement.spacedBy(12.dp),
modifier = Modifier.fillMaxSize()
) {
items(
items = state.photos,
key = { it.image.imageId }
) { photo ->
PhotoCard(
photo = photo,
onImageClick = onImageClick
)
}
}
}
}
}
@Composable
private fun StatItem(
icon: androidx.compose.ui.graphics.vector.ImageVector,
label: String,
value: String
) {
Column(
horizontalAlignment = Alignment.CenterHorizontally,
verticalArrangement = Arrangement.spacedBy(4.dp)
) {
Icon(
icon,
contentDescription = null,
modifier = Modifier.size(24.dp),
tint = MaterialTheme.colorScheme.primary
)
Text(
text = value,
style = MaterialTheme.typography.titleLarge,
fontWeight = FontWeight.Bold
)
Text(
text = label,
style = MaterialTheme.typography.bodySmall,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
}
}
/**
* PhotoCard - CLEAN VERSION: Simple image + person names
*/
@Composable
private fun PhotoCard(
photo: AlbumPhoto,
onImageClick: (String) -> Unit
) {
Card(
modifier = Modifier
.fillMaxWidth()
.aspectRatio(1f)
.clickable { onImageClick(photo.image.imageUri) },
shape = RoundedCornerShape(12.dp)
) {
Box {
// Image
AsyncImage(
model = photo.image.imageUri,
contentDescription = null,
modifier = Modifier.fillMaxSize(),
contentScale = androidx.compose.ui.layout.ContentScale.Crop
)
// Person names overlay (if any)
if (photo.persons.isNotEmpty()) {
Surface(
color = MaterialTheme.colorScheme.surface.copy(alpha = 0.9f),
modifier = Modifier
.align(Alignment.BottomCenter)
.fillMaxWidth()
) {
Text(
text = photo.persons.take(2).joinToString(", ") { it.name },
style = MaterialTheme.typography.bodySmall,
modifier = Modifier.padding(8.dp),
maxLines = 1,
overflow = TextOverflow.Ellipsis,
fontWeight = FontWeight.Medium
)
}
}
}
}
}
/**
* Export Dialog
*/
@Composable
private fun ExportDialog(
albumName: String,
photoCount: Int,
onDismiss: () -> Unit,
onExportToFolder: () -> Unit,
onExportToZip: () -> Unit,
onExportToCollage: () -> Unit
) {
AlertDialog(
onDismissRequest = onDismiss,
icon = { Icon(Icons.Default.FileDownload, null) },
title = { Text("Export Album") },
text = {
Column(
modifier = Modifier.fillMaxWidth(),
verticalArrangement = Arrangement.spacedBy(12.dp)
) {
Text(
"$photoCount photos from \"$albumName\"",
style = MaterialTheme.typography.bodyMedium,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
// Export to Folder
ExportOption(
icon = Icons.Default.Folder,
title = "Export to Folder",
description = "Save all photos to a folder",
onClick = onExportToFolder
)
// Export to Zip
ExportOption(
icon = Icons.Default.FolderZip,
title = "Export as ZIP",
description = "Create a compressed archive",
onClick = onExportToZip
)
// Export to Collage (placeholder)
ExportOption(
icon = Icons.Default.GridView,
title = "Create Collage",
description = "Coming soon!",
onClick = onExportToCollage,
enabled = false
)
}
},
confirmButton = {
TextButton(onClick = onDismiss) {
Text("Cancel")
}
}
)
}
@Composable
private fun ExportOption(
icon: androidx.compose.ui.graphics.vector.ImageVector,
title: String,
description: String,
onClick: () -> Unit,
enabled: Boolean = true
) {
Surface(
modifier = Modifier
.fillMaxWidth()
.clickable(enabled = enabled, onClick = onClick),
shape = RoundedCornerShape(12.dp),
color = if (enabled) {
MaterialTheme.colorScheme.surfaceVariant
} else {
MaterialTheme.colorScheme.surfaceVariant.copy(alpha = 0.5f)
}
) {
Row(
modifier = Modifier.padding(16.dp),
horizontalArrangement = Arrangement.spacedBy(16.dp),
verticalAlignment = Alignment.CenterVertically
) {
Surface(
shape = RoundedCornerShape(8.dp),
color = MaterialTheme.colorScheme.primary.copy(
alpha = if (enabled) 1f else 0.5f
),
modifier = Modifier.size(40.dp)
) {
Box(contentAlignment = Alignment.Center) {
Icon(
icon,
contentDescription = null,
modifier = Modifier.size(24.dp),
tint = MaterialTheme.colorScheme.onPrimary
)
}
}
Column(modifier = Modifier.weight(1f)) {
Text(
title,
style = MaterialTheme.typography.titleMedium,
fontWeight = FontWeight.SemiBold,
color = if (enabled) {
MaterialTheme.colorScheme.onSurface
} else {
MaterialTheme.colorScheme.onSurface.copy(alpha = 0.5f)
}
)
Text(
description,
style = MaterialTheme.typography.bodySmall,
color = MaterialTheme.colorScheme.onSurfaceVariant.copy(
alpha = if (enabled) 1f else 0.5f
)
)
}
if (enabled) {
Icon(
Icons.Default.ChevronRight,
contentDescription = null,
tint = MaterialTheme.colorScheme.onSurfaceVariant
)
}
}
}
}

View File

@@ -0,0 +1,389 @@
package com.placeholder.sherpai2.ui.collections
import androidx.compose.foundation.clickable
import androidx.compose.foundation.layout.*
import androidx.compose.foundation.lazy.grid.*
import androidx.compose.foundation.shape.RoundedCornerShape
import androidx.compose.material.icons.Icons
import androidx.compose.material.icons.filled.*
import androidx.compose.material3.*
import androidx.compose.runtime.*
import androidx.compose.ui.Alignment
import androidx.compose.ui.Modifier
import androidx.compose.ui.graphics.Color
import androidx.compose.ui.text.font.FontWeight
import androidx.compose.ui.text.style.TextOverflow
import androidx.compose.ui.unit.dp
import androidx.hilt.navigation.compose.hiltViewModel
import androidx.lifecycle.compose.collectAsStateWithLifecycle
import coil.compose.AsyncImage
/**
* CollectionsScreen - Main collections list
*
* Features:
* - Grid of collection cards
* - Create new collection button
* - Filter by type (all, smart, static)
* - Collection details on click
*/
@OptIn(ExperimentalMaterial3Api::class)
@Composable
fun CollectionsScreen(
onCollectionClick: (String) -> Unit,
onCreateClick: () -> Unit,
viewModel: CollectionsViewModel = hiltViewModel()
) {
val collections by viewModel.collections.collectAsStateWithLifecycle()
val creationState by viewModel.creationState.collectAsStateWithLifecycle()
Scaffold(
topBar = {
TopAppBar(
title = {
Column {
Text(
"Collections",
style = MaterialTheme.typography.titleLarge,
fontWeight = FontWeight.Bold
)
Text(
viewModel.getCollectionSummary(),
style = MaterialTheme.typography.bodySmall,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
}
},
actions = {
IconButton(onClick = { viewModel.refreshSmartCollections() }) {
Icon(Icons.Default.Refresh, "Refresh smart collections")
}
}
)
},
floatingActionButton = {
ExtendedFloatingActionButton(
onClick = onCreateClick,
icon = { Icon(Icons.Default.Add, null) },
text = { Text("New Collection") }
)
}
) { paddingValues ->
if (collections.isEmpty()) {
EmptyState(
onCreateClick = onCreateClick,
modifier = Modifier
.fillMaxSize()
.padding(paddingValues)
)
} else {
LazyVerticalGrid(
columns = GridCells.Adaptive(160.dp),
modifier = Modifier
.fillMaxSize()
.padding(paddingValues),
contentPadding = PaddingValues(16.dp),
horizontalArrangement = Arrangement.spacedBy(12.dp),
verticalArrangement = Arrangement.spacedBy(12.dp)
) {
items(
items = collections,
key = { it.collectionId }
) { collection ->
CollectionCard(
collection = collection,
onClick = { onCollectionClick(collection.collectionId) },
onPinToggle = { viewModel.togglePinned(collection.collectionId) },
onDelete = { viewModel.deleteCollection(collection.collectionId) }
)
}
}
}
}
// Creation dialog (shown from SearchScreen or other places)
when (val state = creationState) {
is CreationState.SmartFromSearch -> {
CreateCollectionDialog(
title = "Smart Collection",
subtitle = "${state.photoCount} photos matching filters",
onConfirm = { name, description ->
viewModel.createSmartCollection(name, description)
},
onDismiss = { viewModel.cancelCreation() }
)
}
is CreationState.StaticFromImages -> {
CreateCollectionDialog(
title = "Static Collection",
subtitle = "${state.photoCount} photos selected",
onConfirm = { name, description ->
viewModel.createStaticCollection(name, description)
},
onDismiss = { viewModel.cancelCreation() }
)
}
CreationState.None -> { /* No dialog */ }
}
}
@Composable
private fun CollectionCard(
collection: com.placeholder.sherpai2.data.local.entity.CollectionEntity,
onClick: () -> Unit,
onPinToggle: () -> Unit,
onDelete: () -> Unit
) {
var showMenu by remember { mutableStateOf(false) }
Card(
modifier = Modifier
.fillMaxWidth()
.aspectRatio(0.75f)
.clickable(onClick = onClick),
shape = RoundedCornerShape(16.dp)
) {
Box(modifier = Modifier.fillMaxSize()) {
// Cover image or placeholder
if (collection.coverImageUri != null) {
AsyncImage(
model = collection.coverImageUri,
contentDescription = null,
modifier = Modifier.fillMaxSize(),
contentScale = androidx.compose.ui.layout.ContentScale.Crop
)
} else {
// Placeholder
Surface(
modifier = Modifier.fillMaxSize(),
color = MaterialTheme.colorScheme.surfaceVariant
) {
Icon(
Icons.Default.Photo,
contentDescription = null,
modifier = Modifier
.fillMaxSize()
.padding(48.dp),
tint = MaterialTheme.colorScheme.onSurfaceVariant.copy(alpha = 0.3f)
)
}
}
// Gradient overlay for text
Surface(
modifier = Modifier
.fillMaxWidth()
.align(Alignment.BottomCenter),
color = Color.Black.copy(alpha = 0.6f)
) {
Column(
modifier = Modifier.padding(12.dp),
verticalArrangement = Arrangement.spacedBy(4.dp)
) {
Row(
modifier = Modifier.fillMaxWidth(),
horizontalArrangement = Arrangement.SpaceBetween,
verticalAlignment = Alignment.CenterVertically
) {
Text(
collection.name,
style = MaterialTheme.typography.titleMedium,
fontWeight = FontWeight.Bold,
color = Color.White,
maxLines = 1,
overflow = TextOverflow.Ellipsis,
modifier = Modifier.weight(1f)
)
Box {
IconButton(
onClick = { showMenu = true },
modifier = Modifier.size(24.dp)
) {
Icon(
Icons.Default.MoreVert,
null,
tint = Color.White
)
}
DropdownMenu(
expanded = showMenu,
onDismissRequest = { showMenu = false }
) {
DropdownMenuItem(
text = { Text(if (collection.isPinned) "Unpin" else "Pin") },
onClick = {
onPinToggle()
showMenu = false
},
leadingIcon = {
Icon(
if (collection.isPinned) Icons.Default.PushPin else Icons.Default.PushPin,
null
)
}
)
DropdownMenuItem(
text = { Text("Delete") },
onClick = {
onDelete()
showMenu = false
},
leadingIcon = {
Icon(Icons.Default.Delete, null)
}
)
}
}
}
Row(
horizontalArrangement = Arrangement.spacedBy(8.dp),
verticalAlignment = Alignment.CenterVertically
) {
// Type badge
Surface(
color = when (collection.type) {
"SMART" -> Color(0xFF2196F3)
"FAVORITE" -> Color(0xFFF44336)
else -> Color(0xFF4CAF50)
}.copy(alpha = 0.9f),
shape = RoundedCornerShape(4.dp)
) {
Text(
when (collection.type) {
"SMART" -> "Smart"
"FAVORITE" -> "Fav"
else -> "Static"
},
modifier = Modifier.padding(horizontal = 6.dp, vertical = 2.dp),
style = MaterialTheme.typography.labelSmall,
color = Color.White
)
}
// Photo count
Text(
"${collection.photoCount} photos",
style = MaterialTheme.typography.bodySmall,
color = Color.White.copy(alpha = 0.9f)
)
}
}
}
// Pinned indicator
if (collection.isPinned) {
Icon(
Icons.Default.PushPin,
contentDescription = "Pinned",
modifier = Modifier
.align(Alignment.TopEnd)
.padding(8.dp)
.size(20.dp),
tint = Color.White
)
}
}
}
}
@Composable
private fun EmptyState(
onCreateClick: () -> Unit,
modifier: Modifier = Modifier
) {
Box(
modifier = modifier,
contentAlignment = Alignment.Center
) {
Column(
horizontalAlignment = Alignment.CenterHorizontally,
verticalArrangement = Arrangement.spacedBy(16.dp)
) {
Icon(
Icons.Default.Collections,
contentDescription = null,
modifier = Modifier.size(72.dp),
tint = MaterialTheme.colorScheme.onSurfaceVariant.copy(alpha = 0.6f)
)
Text(
"No Collections Yet",
style = MaterialTheme.typography.titleLarge,
fontWeight = FontWeight.Bold
)
Text(
"Create collections from searches or manually select photos",
style = MaterialTheme.typography.bodyMedium,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
Button(onClick = onCreateClick) {
Icon(Icons.Default.Add, null, Modifier.size(18.dp))
Spacer(Modifier.width(8.dp))
Text("Create Collection")
}
}
}
}
@Composable
private fun CreateCollectionDialog(
title: String,
subtitle: String,
onConfirm: (name: String, description: String?) -> Unit,
onDismiss: () -> Unit
) {
var name by remember { mutableStateOf("") }
var description by remember { mutableStateOf("") }
AlertDialog(
onDismissRequest = onDismiss,
icon = { Icon(Icons.Default.Collections, null) },
title = {
Column {
Text(title)
Text(
subtitle,
style = MaterialTheme.typography.bodySmall,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
}
},
text = {
Column(verticalArrangement = Arrangement.spacedBy(12.dp)) {
OutlinedTextField(
value = name,
onValueChange = { name = it },
label = { Text("Collection Name") },
singleLine = true,
modifier = Modifier.fillMaxWidth()
)
OutlinedTextField(
value = description,
onValueChange = { description = it },
label = { Text("Description (optional)") },
maxLines = 3,
modifier = Modifier.fillMaxWidth()
)
}
},
confirmButton = {
Button(
onClick = {
if (name.isNotBlank()) {
onConfirm(name.trim(), description.trim().ifBlank { null })
}
},
enabled = name.isNotBlank()
) {
Text("Create")
}
},
dismissButton = {
TextButton(onClick = onDismiss) {
Text("Cancel")
}
}
)
}

View File

@@ -0,0 +1,159 @@
package com.placeholder.sherpai2.ui.collections
import androidx.lifecycle.ViewModel
import androidx.lifecycle.viewModelScope
import com.placeholder.sherpai2.data.local.entity.CollectionEntity
import com.placeholder.sherpai2.data.repository.CollectionRepository
import com.placeholder.sherpai2.ui.search.DateRange
import dagger.hilt.android.lifecycle.HiltViewModel
import kotlinx.coroutines.flow.*
import kotlinx.coroutines.launch
import javax.inject.Inject
/**
* CollectionsViewModel - Manages collections list and creation
*/
@HiltViewModel
class CollectionsViewModel @Inject constructor(
private val collectionRepository: CollectionRepository
) : ViewModel() {
// All collections
val collections: StateFlow<List<CollectionEntity>> = collectionRepository
.getAllCollections()
.stateIn(
scope = viewModelScope,
started = SharingStarted.WhileSubscribed(5000),
initialValue = emptyList()
)
// UI state for creation dialog
private val _creationState = MutableStateFlow<CreationState>(CreationState.None)
val creationState: StateFlow<CreationState> = _creationState.asStateFlow()
// ==========================================
// COLLECTION CREATION
// ==========================================
fun startSmartCollectionFromSearch(
includedPeople: Set<String>,
excludedPeople: Set<String>,
includedTags: Set<String>,
excludedTags: Set<String>,
dateRange: DateRange,
photoCount: Int
) {
_creationState.value = CreationState.SmartFromSearch(
includedPeople = includedPeople,
excludedPeople = excludedPeople,
includedTags = includedTags,
excludedTags = excludedTags,
dateRange = dateRange,
photoCount = photoCount
)
}
fun startStaticCollectionFromImages(imageIds: List<String>) {
_creationState.value = CreationState.StaticFromImages(
imageIds = imageIds,
photoCount = imageIds.size
)
}
fun cancelCreation() {
_creationState.value = CreationState.None
}
fun createSmartCollection(name: String, description: String?) {
val state = _creationState.value as? CreationState.SmartFromSearch ?: return
viewModelScope.launch {
collectionRepository.createSmartCollection(
name = name,
description = description,
includedPeople = state.includedPeople,
excludedPeople = state.excludedPeople,
includedTags = state.includedTags,
excludedTags = state.excludedTags,
dateRange = state.dateRange
)
_creationState.value = CreationState.None
}
}
fun createStaticCollection(name: String, description: String?) {
val state = _creationState.value as? CreationState.StaticFromImages ?: return
viewModelScope.launch {
collectionRepository.createStaticCollection(
name = name,
description = description,
imageIds = state.imageIds
)
_creationState.value = CreationState.None
}
}
// ==========================================
// COLLECTION MANAGEMENT
// ==========================================
fun deleteCollection(collectionId: String) {
viewModelScope.launch {
collectionRepository.deleteCollection(collectionId)
}
}
fun togglePinned(collectionId: String) {
viewModelScope.launch {
collectionRepository.togglePinned(collectionId)
}
}
fun refreshSmartCollections() {
viewModelScope.launch {
collectionRepository.evaluateAllSmartCollections()
}
}
// ==========================================
// STATISTICS
// ==========================================
fun getCollectionSummary(): String {
val count = collections.value.size
val smartCount = collections.value.count { it.type == "SMART" }
val staticCount = collections.value.count { it.type == "STATIC" }
return when {
count == 0 -> "No collections yet"
smartCount > 0 && staticCount > 0 -> "$smartCount smart • $staticCount static"
smartCount > 0 -> "$smartCount smart collections"
staticCount > 0 -> "$staticCount static collections"
else -> "$count collections"
}
}
}
/**
* Creation state for dialogs
*/
sealed class CreationState {
object None : CreationState()
data class SmartFromSearch(
val includedPeople: Set<String>,
val excludedPeople: Set<String>,
val includedTags: Set<String>,
val excludedTags: Set<String>,
val dateRange: DateRange,
val photoCount: Int
) : CreationState()
data class StaticFromImages(
val imageIds: List<String>,
val photoCount: Int
) : CreationState()
}

View File

@@ -0,0 +1,162 @@
package com.placeholder.sherpai2.ui.devscreens
import androidx.compose.foundation.background
import androidx.compose.foundation.layout.*
import androidx.compose.foundation.shape.RoundedCornerShape
import androidx.compose.material.icons.Icons
import androidx.compose.material.icons.filled.*
import androidx.compose.material3.*
import androidx.compose.runtime.Composable
import androidx.compose.ui.Alignment
import androidx.compose.ui.Modifier
import androidx.compose.ui.graphics.Brush
import androidx.compose.ui.text.font.FontWeight
import androidx.compose.ui.text.style.TextAlign
import androidx.compose.ui.unit.dp
/**
* Beautiful placeholder screen for features under development
*
* Shows:
* - Feature name
* - Description
* - "Coming Soon" indicator
* - Consistent styling with rest of app
*/
@Composable
fun DummyScreen(
title: String,
subtitle: String = "This feature is under development"
) {
Box(
modifier = Modifier
.fillMaxSize()
.background(
Brush.verticalGradient(
colors = listOf(
MaterialTheme.colorScheme.surface,
MaterialTheme.colorScheme.surfaceVariant.copy(alpha = 0.3f)
)
)
),
contentAlignment = Alignment.Center
) {
Column(
horizontalAlignment = Alignment.CenterHorizontally,
verticalArrangement = Arrangement.spacedBy(24.dp),
modifier = Modifier.padding(48.dp)
) {
// Icon badge
Surface(
modifier = Modifier.size(96.dp),
shape = RoundedCornerShape(24.dp),
color = MaterialTheme.colorScheme.primaryContainer,
shadowElevation = 8.dp
) {
Box(contentAlignment = Alignment.Center) {
Icon(
Icons.Default.Construction,
contentDescription = null,
modifier = Modifier.size(48.dp),
tint = MaterialTheme.colorScheme.primary
)
}
}
Spacer(modifier = Modifier.height(8.dp))
// Title
Text(
text = title,
style = MaterialTheme.typography.headlineMedium,
fontWeight = FontWeight.Bold,
textAlign = TextAlign.Center
)
// Subtitle
Text(
text = subtitle,
style = MaterialTheme.typography.bodyLarge,
color = MaterialTheme.colorScheme.onSurfaceVariant,
textAlign = TextAlign.Center,
modifier = Modifier.padding(horizontal = 24.dp)
)
Spacer(modifier = Modifier.height(8.dp))
// Coming soon badge
Surface(
shape = RoundedCornerShape(16.dp),
color = MaterialTheme.colorScheme.tertiaryContainer,
shadowElevation = 2.dp
) {
Row(
modifier = Modifier.padding(horizontal = 20.dp, vertical = 12.dp),
horizontalArrangement = Arrangement.spacedBy(8.dp),
verticalAlignment = Alignment.CenterVertically
) {
Icon(
Icons.Default.Schedule,
contentDescription = null,
modifier = Modifier.size(20.dp),
tint = MaterialTheme.colorScheme.onTertiaryContainer
)
Text(
text = "Coming Soon",
style = MaterialTheme.typography.labelLarge,
fontWeight = FontWeight.SemiBold,
color = MaterialTheme.colorScheme.onTertiaryContainer
)
}
}
Spacer(modifier = Modifier.height(24.dp))
// Feature preview card
Card(
modifier = Modifier.fillMaxWidth(0.8f),
colors = CardDefaults.cardColors(
containerColor = MaterialTheme.colorScheme.surfaceVariant.copy(alpha = 0.5f)
),
shape = RoundedCornerShape(16.dp)
) {
Column(
modifier = Modifier.padding(20.dp),
verticalArrangement = Arrangement.spacedBy(12.dp)
) {
Text(
text = "What's planned:",
style = MaterialTheme.typography.titleSmall,
fontWeight = FontWeight.Bold
)
FeatureItem("Full implementation")
FeatureItem("Beautiful UI design")
FeatureItem("Smooth animations")
FeatureItem("Production-ready code")
}
}
}
}
}
@Composable
private fun FeatureItem(text: String) {
Row(
horizontalArrangement = Arrangement.spacedBy(8.dp),
verticalAlignment = Alignment.CenterVertically
) {
Icon(
Icons.Default.CheckCircle,
contentDescription = null,
modifier = Modifier.size(16.dp),
tint = MaterialTheme.colorScheme.primary
)
Text(
text = text,
style = MaterialTheme.typography.bodyMedium,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
}
}

View File

@@ -0,0 +1,687 @@
package com.placeholder.sherpai2.ui.discover
import android.graphics.BitmapFactory
import android.net.Uri
import androidx.compose.foundation.Image
import androidx.compose.foundation.background
import androidx.compose.foundation.clickable
import androidx.compose.foundation.layout.*
import androidx.compose.foundation.lazy.LazyColumn
import androidx.compose.foundation.lazy.grid.GridCells
import androidx.compose.foundation.lazy.grid.LazyVerticalGrid
import androidx.compose.foundation.lazy.grid.items
import androidx.compose.foundation.lazy.items
import androidx.compose.foundation.shape.CircleShape
import androidx.compose.foundation.shape.RoundedCornerShape
import androidx.compose.material.icons.Icons
import androidx.compose.material.icons.filled.*
import androidx.compose.material3.*
import androidx.compose.runtime.*
import androidx.compose.ui.Alignment
import androidx.compose.ui.Modifier
import androidx.compose.ui.draw.clip
import androidx.compose.ui.graphics.asImageBitmap
import androidx.compose.ui.layout.ContentScale
import androidx.compose.ui.platform.LocalContext
import androidx.compose.ui.text.font.FontWeight
import androidx.compose.ui.text.style.TextAlign
import androidx.compose.ui.unit.dp
import androidx.hilt.navigation.compose.hiltViewModel
import androidx.lifecycle.compose.collectAsStateWithLifecycle
import com.placeholder.sherpai2.domain.clustering.AgeEstimate
import com.placeholder.sherpai2.domain.clustering.FaceCluster
import java.text.SimpleDateFormat
import java.util.*
/**
* DiscoverPeopleScreen - Beautiful auto-clustering UI
*
* FLOW:
* 1. Hero CTA: "Discover People in Your Photos"
* 2. Auto-clustering progress (2-5 min)
* 3. Grid of discovered people
* 4. Tap cluster → Name person + metadata
* 5. Background deep scan starts
*/
@OptIn(ExperimentalMaterial3Api::class)
@Composable
fun DiscoverPeopleScreen(
viewModel: DiscoverPeopleViewModel = hiltViewModel()
) {
val uiState by viewModel.uiState.collectAsStateWithLifecycle()
// NO SCAFFOLD - MainScreen already has TopAppBar
Box(modifier = Modifier.fillMaxSize()) {
when (val state = uiState) {
is DiscoverUiState.Idle -> IdleScreen(
onStartDiscovery = { viewModel.startDiscovery() }
)
is DiscoverUiState.Clustering -> ClusteringProgressScreen(
progress = state.progress,
total = state.total,
message = state.message
)
is DiscoverUiState.NamingReady -> ClusterGridScreen(
result = state.result,
onClusterClick = { cluster ->
viewModel.selectCluster(cluster)
}
)
is DiscoverUiState.NamingCluster -> NamingDialog(
cluster = state.selectedCluster,
suggestedSiblings = state.suggestedSiblings,
onConfirm = { name, dob, isChild, siblings ->
viewModel.confirmClusterName(
cluster = state.selectedCluster,
name = name,
dateOfBirth = dob,
isChild = isChild,
selectedSiblings = siblings
)
},
onDismiss = { viewModel.cancelNaming() }
)
is DiscoverUiState.NoPeopleFound -> EmptyStateScreen(
message = state.message
)
is DiscoverUiState.Error -> ErrorScreen(
message = state.message,
onRetry = { viewModel.startDiscovery() }
)
}
}
}
/**
* Idle state - Hero CTA to start discovery
*/
@Composable
fun IdleScreen(
onStartDiscovery: () -> Unit
) {
Column(
modifier = Modifier
.fillMaxSize()
.padding(32.dp),
horizontalAlignment = Alignment.CenterHorizontally,
verticalArrangement = Arrangement.Center
) {
Icon(
imageVector = Icons.Default.AutoAwesome,
contentDescription = null,
modifier = Modifier.size(120.dp),
tint = MaterialTheme.colorScheme.primary
)
Spacer(Modifier.height(24.dp))
Text(
text = "Discover People",
style = MaterialTheme.typography.headlineLarge,
fontWeight = FontWeight.Bold,
textAlign = TextAlign.Center
)
Spacer(Modifier.height(16.dp))
Text(
text = "Let AI automatically find and group faces in your photos. " +
"You'll name them, and we'll tag all their photos.",
style = MaterialTheme.typography.bodyLarge,
textAlign = TextAlign.Center,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
Spacer(Modifier.height(32.dp))
Button(
onClick = onStartDiscovery,
modifier = Modifier
.fillMaxWidth()
.height(56.dp),
colors = ButtonDefaults.buttonColors(
containerColor = MaterialTheme.colorScheme.primary
)
) {
Icon(
imageVector = Icons.Default.AutoAwesome,
contentDescription = null,
modifier = Modifier.size(24.dp)
)
Spacer(Modifier.width(8.dp))
Text(
text = "Start Discovery",
style = MaterialTheme.typography.titleMedium,
fontWeight = FontWeight.Bold
)
}
Spacer(Modifier.height(16.dp))
Card(
modifier = Modifier.fillMaxWidth(),
colors = CardDefaults.cardColors(
containerColor = MaterialTheme.colorScheme.surfaceVariant
)
) {
Column(
modifier = Modifier.padding(16.dp),
verticalArrangement = Arrangement.spacedBy(8.dp)
) {
InfoRow(Icons.Default.Speed, "Fast: Analyzes ~1000 photos in 2-5 minutes")
InfoRow(Icons.Default.Security, "Private: Everything stays on your device")
InfoRow(Icons.Default.AutoAwesome, "Smart: Groups faces automatically")
}
}
}
}
@Composable
fun InfoRow(icon: androidx.compose.ui.graphics.vector.ImageVector, text: String) {
Row(
verticalAlignment = Alignment.CenterVertically,
horizontalArrangement = Arrangement.spacedBy(12.dp)
) {
Icon(
imageVector = icon,
contentDescription = null,
tint = MaterialTheme.colorScheme.primary,
modifier = Modifier.size(20.dp)
)
Text(
text = text,
style = MaterialTheme.typography.bodyMedium
)
}
}
/**
* Clustering progress screen
*/
@Composable
fun ClusteringProgressScreen(
progress: Int,
total: Int,
message: String
) {
Column(
modifier = Modifier
.fillMaxSize()
.padding(32.dp),
horizontalAlignment = Alignment.CenterHorizontally,
verticalArrangement = Arrangement.Center
) {
CircularProgressIndicator(
modifier = Modifier.size(80.dp),
strokeWidth = 6.dp
)
Spacer(Modifier.height(32.dp))
Text(
text = "Discovering People...",
style = MaterialTheme.typography.headlineSmall,
fontWeight = FontWeight.Bold
)
Spacer(Modifier.height(16.dp))
LinearProgressIndicator(
progress = { if (total > 0) progress.toFloat() / total.toFloat() else 0f },
modifier = Modifier.fillMaxWidth(),
)
Spacer(Modifier.height(8.dp))
Text(
text = message,
style = MaterialTheme.typography.bodyMedium,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
Spacer(Modifier.height(24.dp))
Text(
text = "This will take 2-5 minutes. You can leave and come back later.",
style = MaterialTheme.typography.bodySmall,
textAlign = TextAlign.Center,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
}
}
/**
* Grid of discovered clusters
*/
@Composable
fun ClusterGridScreen(
result: com.placeholder.sherpai2.domain.clustering.ClusteringResult,
onClusterClick: (FaceCluster) -> Unit
) {
Column(
modifier = Modifier
.fillMaxSize()
.padding(16.dp)
) {
Text(
text = "Found ${result.clusters.size} People",
style = MaterialTheme.typography.headlineSmall,
fontWeight = FontWeight.Bold
)
Spacer(Modifier.height(8.dp))
Text(
text = "Tap to name each person. We'll then tag all their photos.",
style = MaterialTheme.typography.bodyMedium,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
Spacer(Modifier.height(16.dp))
LazyVerticalGrid(
columns = GridCells.Fixed(2),
horizontalArrangement = Arrangement.spacedBy(12.dp),
verticalArrangement = Arrangement.spacedBy(12.dp)
) {
items(result.clusters) { cluster ->
ClusterCard(
cluster = cluster,
onClick = { onClusterClick(cluster) }
)
}
}
}
}
/**
* Single cluster card
*/
@Composable
fun ClusterCard(
cluster: FaceCluster,
onClick: () -> Unit
) {
val context = LocalContext.current
Card(
modifier = Modifier
.fillMaxWidth()
.clickable(onClick = onClick),
elevation = CardDefaults.cardElevation(defaultElevation = 2.dp)
) {
Column {
// Face grid (2x3)
LazyVerticalGrid(
columns = GridCells.Fixed(3),
modifier = Modifier.height(180.dp),
userScrollEnabled = false
) {
items(cluster.representativeFaces.take(6)) { face ->
val bitmap = remember(face.imageUri) {
try {
context.contentResolver.openInputStream(Uri.parse(face.imageUri))?.use {
BitmapFactory.decodeStream(it)
}
} catch (e: Exception) {
null
}
}
if (bitmap != null) {
Image(
bitmap = bitmap.asImageBitmap(),
contentDescription = null,
modifier = Modifier
.fillMaxWidth()
.aspectRatio(1f),
contentScale = ContentScale.Crop
)
} else {
Box(
modifier = Modifier
.fillMaxWidth()
.aspectRatio(1f)
.background(MaterialTheme.colorScheme.surfaceVariant),
contentAlignment = Alignment.Center
) {
Icon(
imageVector = Icons.Default.Person,
contentDescription = null,
tint = MaterialTheme.colorScheme.onSurfaceVariant
)
}
}
}
}
// Info
Column(
modifier = Modifier.padding(12.dp)
) {
Row(
modifier = Modifier.fillMaxWidth(),
horizontalArrangement = Arrangement.SpaceBetween,
verticalAlignment = Alignment.CenterVertically
) {
Text(
text = "${cluster.photoCount} photos",
style = MaterialTheme.typography.titleMedium,
fontWeight = FontWeight.Bold
)
if (cluster.estimatedAge == AgeEstimate.CHILD) {
Surface(
shape = RoundedCornerShape(12.dp),
color = MaterialTheme.colorScheme.primaryContainer
) {
Text(
text = "Child",
modifier = Modifier.padding(horizontal = 8.dp, vertical = 4.dp),
style = MaterialTheme.typography.labelSmall,
color = MaterialTheme.colorScheme.onPrimaryContainer
)
}
}
}
if (cluster.potentialSiblings.isNotEmpty()) {
Spacer(Modifier.height(4.dp))
Text(
text = "Appears with ${cluster.potentialSiblings.size} other person(s)",
style = MaterialTheme.typography.bodySmall,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
}
}
}
}
}
/**
* Naming dialog
*/
@Composable
fun NamingDialog(
cluster: FaceCluster,
suggestedSiblings: List<FaceCluster>,
onConfirm: (String, Long?, Boolean, List<Int>) -> Unit,
onDismiss: () -> Unit
) {
var name by remember { mutableStateOf("") }
var isChild by remember { mutableStateOf(cluster.estimatedAge == AgeEstimate.CHILD) }
var dateOfBirth by remember { mutableStateOf<Long?>(null) }
var selectedSiblings by remember { mutableStateOf<Set<Int>>(emptySet()) }
var showDatePicker by remember { mutableStateOf(false) }
val context = LocalContext.current
// Date picker dialog
if (showDatePicker) {
val calendar = java.util.Calendar.getInstance()
if (dateOfBirth != null) {
calendar.timeInMillis = dateOfBirth!!
}
val datePickerDialog = android.app.DatePickerDialog(
context,
{ _, year, month, dayOfMonth ->
val cal = java.util.Calendar.getInstance()
cal.set(year, month, dayOfMonth)
dateOfBirth = cal.timeInMillis
showDatePicker = false
},
calendar.get(java.util.Calendar.YEAR),
calendar.get(java.util.Calendar.MONTH),
calendar.get(java.util.Calendar.DAY_OF_MONTH)
)
datePickerDialog.setOnDismissListener {
showDatePicker = false
}
DisposableEffect(Unit) {
datePickerDialog.show()
onDispose {
datePickerDialog.dismiss()
}
}
}
AlertDialog(
onDismissRequest = onDismiss,
title = {
Text("Name This Person")
},
text = {
Column(
verticalArrangement = Arrangement.spacedBy(16.dp)
) {
// FACE PREVIEW - Show 6 representative faces
Text(
text = "${cluster.photoCount} photos found",
style = MaterialTheme.typography.bodyMedium,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
LazyVerticalGrid(
columns = GridCells.Fixed(3),
modifier = Modifier.height(180.dp),
horizontalArrangement = Arrangement.spacedBy(4.dp),
verticalArrangement = Arrangement.spacedBy(4.dp)
) {
items(cluster.representativeFaces.take(6)) { face ->
val bitmap = remember(face.imageUri) {
try {
context.contentResolver.openInputStream(Uri.parse(face.imageUri))?.use {
BitmapFactory.decodeStream(it)
}
} catch (e: Exception) {
null
}
}
if (bitmap != null) {
Image(
bitmap = bitmap.asImageBitmap(),
contentDescription = null,
modifier = Modifier
.fillMaxWidth()
.aspectRatio(1f)
.clip(RoundedCornerShape(8.dp)),
contentScale = ContentScale.Crop
)
} else {
Box(
modifier = Modifier
.fillMaxWidth()
.aspectRatio(1f)
.clip(RoundedCornerShape(8.dp))
.background(MaterialTheme.colorScheme.surfaceVariant),
contentAlignment = Alignment.Center
) {
Icon(
imageVector = Icons.Default.Person,
contentDescription = null,
tint = MaterialTheme.colorScheme.onSurfaceVariant
)
}
}
}
}
HorizontalDivider()
// Name input
OutlinedTextField(
value = name,
onValueChange = { name = it },
label = { Text("Name") },
singleLine = true,
modifier = Modifier.fillMaxWidth()
)
// Is child toggle
Row(
modifier = Modifier.fillMaxWidth(),
horizontalArrangement = Arrangement.SpaceBetween,
verticalAlignment = Alignment.CenterVertically
) {
Text("This person is a child")
Switch(
checked = isChild,
onCheckedChange = { isChild = it }
)
}
// Date of birth (if child)
if (isChild) {
OutlinedButton(
onClick = { showDatePicker = true },
modifier = Modifier.fillMaxWidth()
) {
Icon(Icons.Default.CalendarToday, null)
Spacer(Modifier.width(8.dp))
Text(
if (dateOfBirth != null) {
SimpleDateFormat("MMM dd, yyyy", Locale.getDefault())
.format(Date(dateOfBirth!!))
} else {
"Set Date of Birth"
}
)
}
}
// Suggested siblings
if (suggestedSiblings.isNotEmpty()) {
Text(
"Appears with these people (select siblings):",
style = MaterialTheme.typography.labelMedium
)
suggestedSiblings.forEach { sibling ->
Row(
modifier = Modifier.fillMaxWidth(),
verticalAlignment = Alignment.CenterVertically
) {
Checkbox(
checked = sibling.clusterId in selectedSiblings,
onCheckedChange = { checked ->
selectedSiblings = if (checked) {
selectedSiblings + sibling.clusterId
} else {
selectedSiblings - sibling.clusterId
}
}
)
Text("Person ${sibling.clusterId + 1} (${sibling.photoCount} photos)")
}
}
}
}
},
confirmButton = {
TextButton(
onClick = {
onConfirm(
name,
dateOfBirth,
isChild,
selectedSiblings.toList()
)
},
enabled = name.isNotBlank()
) {
Text("Save & Train")
}
},
dismissButton = {
TextButton(onClick = onDismiss) {
Text("Cancel")
}
}
)
// TODO: Add DatePickerDialog when showDatePicker is true
}
/**
* Empty state screen
*/
@Composable
fun EmptyStateScreen(message: String) {
Column(
modifier = Modifier
.fillMaxSize()
.padding(32.dp),
horizontalAlignment = Alignment.CenterHorizontally,
verticalArrangement = Arrangement.Center
) {
Icon(
imageVector = Icons.Default.PersonOff,
contentDescription = null,
modifier = Modifier.size(80.dp),
tint = MaterialTheme.colorScheme.onSurfaceVariant
)
Spacer(Modifier.height(16.dp))
Text(
text = message,
style = MaterialTheme.typography.bodyLarge,
textAlign = TextAlign.Center
)
}
}
/**
* Error screen
*/
@Composable
fun ErrorScreen(
message: String,
onRetry: () -> Unit
) {
Column(
modifier = Modifier
.fillMaxSize()
.padding(32.dp),
horizontalAlignment = Alignment.CenterHorizontally,
verticalArrangement = Arrangement.Center
) {
Icon(
imageVector = Icons.Default.Error,
contentDescription = null,
modifier = Modifier.size(80.dp),
tint = MaterialTheme.colorScheme.error
)
Spacer(Modifier.height(16.dp))
Text(
text = "Oops!",
style = MaterialTheme.typography.headlineMedium,
fontWeight = FontWeight.Bold
)
Spacer(Modifier.height(8.dp))
Text(
text = message,
style = MaterialTheme.typography.bodyLarge,
textAlign = TextAlign.Center,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
Spacer(Modifier.height(24.dp))
Button(onClick = onRetry) {
Text("Try Again")
}
}
}

View File

@@ -0,0 +1,222 @@
package com.placeholder.sherpai2.ui.discover
import androidx.lifecycle.ViewModel
import androidx.lifecycle.viewModelScope
import com.placeholder.sherpai2.domain.clustering.ClusteringResult
import com.placeholder.sherpai2.domain.clustering.FaceCluster
import com.placeholder.sherpai2.domain.clustering.FaceClusteringService
import com.placeholder.sherpai2.domain.training.ClusterTrainingService
import dagger.hilt.android.lifecycle.HiltViewModel
import kotlinx.coroutines.flow.MutableStateFlow
import kotlinx.coroutines.flow.StateFlow
import kotlinx.coroutines.flow.asStateFlow
import kotlinx.coroutines.launch
import javax.inject.Inject
/**
* DiscoverPeopleViewModel - Manages auto-clustering and naming flow
*
* PHASE 2: Now includes multi-centroid training from clusters
*
* STATE FLOW:
* 1. Idle → User taps "Discover People"
* 2. Clustering → Auto-analyzing faces (2-5 min)
* 3. NamingReady → Shows clusters, user names them
* 4. Training → Creating multi-centroid face model
* 5. Complete → Ready to scan library
*/
@HiltViewModel
class DiscoverPeopleViewModel @Inject constructor(
private val clusteringService: FaceClusteringService,
private val trainingService: ClusterTrainingService
) : ViewModel() {
private val _uiState = MutableStateFlow<DiscoverUiState>(DiscoverUiState.Idle)
val uiState: StateFlow<DiscoverUiState> = _uiState.asStateFlow()
// Track which clusters have been named
private val namedClusterIds = mutableSetOf<Int>()
/**
* Start auto-clustering process
*/
fun startDiscovery() {
viewModelScope.launch {
try {
// Clear named clusters for new discovery
namedClusterIds.clear()
_uiState.value = DiscoverUiState.Clustering(0, 100, "Starting...")
val result = clusteringService.discoverPeople(
onProgress = { current, total, message ->
_uiState.value = DiscoverUiState.Clustering(current, total, message)
}
)
// Check for errors
if (result.errorMessage != null) {
_uiState.value = DiscoverUiState.Error(result.errorMessage)
return@launch
}
if (result.clusters.isEmpty()) {
_uiState.value = DiscoverUiState.NoPeopleFound(
"No faces found in your library. Make sure face detection cache is populated."
)
} else {
_uiState.value = DiscoverUiState.NamingReady(result)
}
} catch (e: Exception) {
_uiState.value = DiscoverUiState.Error(
e.message ?: "Failed to discover people"
)
}
}
}
/**
* User selected a cluster to name
*/
fun selectCluster(cluster: FaceCluster) {
val currentState = _uiState.value
if (currentState is DiscoverUiState.NamingReady) {
_uiState.value = DiscoverUiState.NamingCluster(
result = currentState.result,
selectedCluster = cluster,
suggestedSiblings = currentState.result.clusters.filter {
it.clusterId in cluster.potentialSiblings
}
)
}
}
/**
* User confirmed name and metadata for a cluster
*
* CREATES:
* 1. PersonEntity with all metadata (name, DOB, siblings)
* 2. Multi-centroid FaceModelEntity (temporal tracking for children)
* 3. Removes cluster from display
*/
fun confirmClusterName(
cluster: FaceCluster,
name: String,
dateOfBirth: Long?,
isChild: Boolean,
selectedSiblings: List<Int>
) {
viewModelScope.launch {
try {
val currentState = _uiState.value
if (currentState !is DiscoverUiState.NamingCluster) return@launch
// Train person from cluster
val personId = trainingService.trainFromCluster(
cluster = cluster,
name = name,
dateOfBirth = dateOfBirth,
isChild = isChild,
siblingClusterIds = selectedSiblings,
onProgress = { current, total, message ->
_uiState.value = DiscoverUiState.Clustering(current, total, message)
}
)
// Mark cluster as named
namedClusterIds.add(cluster.clusterId)
// Filter out named clusters
val remainingClusters = currentState.result.clusters
.filter { it.clusterId !in namedClusterIds }
if (remainingClusters.isEmpty()) {
// All clusters named! Show success
_uiState.value = DiscoverUiState.NoPeopleFound(
"All people have been named! 🎉\n\nGo to 'People' to see your trained models."
)
} else {
// Return to naming screen with remaining clusters
_uiState.value = DiscoverUiState.NamingReady(
result = currentState.result.copy(clusters = remainingClusters)
)
}
} catch (e: Exception) {
_uiState.value = DiscoverUiState.Error(
e.message ?: "Failed to create person: ${e.message}"
)
}
}
}
/**
* Cancel naming and go back to cluster list
*/
fun cancelNaming() {
val currentState = _uiState.value
if (currentState is DiscoverUiState.NamingCluster) {
_uiState.value = DiscoverUiState.NamingReady(
result = currentState.result
)
}
}
/**
* Reset to idle state
*/
fun reset() {
_uiState.value = DiscoverUiState.Idle
}
}
/**
* UI States for Discover People flow
*/
sealed class DiscoverUiState {
/**
* Initial state - user hasn't started discovery
*/
object Idle : DiscoverUiState()
/**
* Auto-clustering in progress
*/
data class Clustering(
val progress: Int,
val total: Int,
val message: String
) : DiscoverUiState()
/**
* Clustering complete, ready for user to name people
*/
data class NamingReady(
val result: ClusteringResult
) : DiscoverUiState()
/**
* User is naming a specific cluster
*/
data class NamingCluster(
val result: ClusteringResult,
val selectedCluster: FaceCluster,
val suggestedSiblings: List<FaceCluster>
) : DiscoverUiState()
/**
* No people found in library
*/
data class NoPeopleFound(
val message: String
) : DiscoverUiState()
/**
* Error occurred
*/
data class Error(
val message: String
) : DiscoverUiState()
}

View File

@@ -0,0 +1,489 @@
package com.placeholder.sherpai2.ui.explore
import androidx.compose.foundation.background
import androidx.compose.foundation.clickable
import androidx.compose.foundation.layout.*
import androidx.compose.foundation.lazy.LazyColumn
import androidx.compose.foundation.lazy.LazyRow
import androidx.compose.foundation.lazy.items
import androidx.compose.foundation.shape.CircleShape
import androidx.compose.foundation.shape.RoundedCornerShape
import androidx.compose.material.icons.Icons
import androidx.compose.material.icons.filled.*
import androidx.compose.material3.*
import androidx.compose.runtime.*
import androidx.compose.ui.Alignment
import androidx.compose.ui.Modifier
import androidx.compose.ui.draw.clip
import androidx.compose.ui.graphics.Brush
import androidx.compose.ui.graphics.Color
import androidx.compose.ui.graphics.vector.ImageVector
import androidx.compose.ui.text.font.FontWeight
import androidx.compose.ui.unit.dp
import androidx.hilt.navigation.compose.hiltViewModel
/**
* CLEANED ExploreScreen - No gradient header banner
*
* Removed:
* - Gradient header box (lines 46-75) that created banner effect
* - "Explore" title (MainScreen shows it)
*
* Features:
* - Rectangular album cards (compact)
* - Stories section (recent highlights)
* - Clickable navigation to AlbumViewScreen
* - Beautiful gradients and icons
* - Mobile-friendly scrolling
*/
@Composable
fun ExploreScreen(
onAlbumClick: (albumType: String, albumId: String) -> Unit,
viewModel: ExploreViewModel = hiltViewModel(),
modifier: Modifier = Modifier
) {
val uiState by viewModel.uiState.collectAsState()
Box(modifier = modifier.fillMaxSize()) {
when (val state = uiState) {
is ExploreViewModel.ExploreUiState.Loading -> {
Box(
modifier = Modifier.fillMaxSize(),
contentAlignment = Alignment.Center
) {
CircularProgressIndicator()
}
}
is ExploreViewModel.ExploreUiState.Success -> {
if (state.smartAlbums.isEmpty()) {
EmptyExploreView()
} else {
ExploreContent(
smartAlbums = state.smartAlbums,
onAlbumClick = onAlbumClick
)
}
}
is ExploreViewModel.ExploreUiState.Error -> {
Box(
modifier = Modifier.fillMaxSize(),
contentAlignment = Alignment.Center
) {
Column(
horizontalAlignment = Alignment.CenterHorizontally,
verticalArrangement = Arrangement.spacedBy(12.dp),
modifier = Modifier.padding(32.dp)
) {
Icon(
Icons.Default.Error,
contentDescription = null,
modifier = Modifier.size(64.dp),
tint = MaterialTheme.colorScheme.error
)
Text(
text = "Error Loading Albums",
style = MaterialTheme.typography.titleLarge,
fontWeight = FontWeight.Bold
)
Text(
text = state.message,
style = MaterialTheme.typography.bodyMedium,
color = MaterialTheme.colorScheme.onSurfaceVariant,
textAlign = androidx.compose.ui.text.style.TextAlign.Center
)
}
}
}
}
}
}
/**
* Main content - scrollable album sections
*/
@Composable
private fun ExploreContent(
smartAlbums: List<SmartAlbum>,
onAlbumClick: (albumType: String, albumId: String) -> Unit
) {
LazyColumn(
modifier = Modifier.fillMaxSize(),
contentPadding = PaddingValues(vertical = 16.dp),
verticalArrangement = Arrangement.spacedBy(24.dp)
) {
// Stories Section (Recent Highlights)
item {
val storyAlbums = smartAlbums.filter { it.imageCount > 0 }.take(10)
if (storyAlbums.isNotEmpty()) {
StoriesSection(
albums = storyAlbums,
onAlbumClick = onAlbumClick
)
}
}
// Time-based Albums
val timeAlbums = smartAlbums.filterIsInstance<SmartAlbum.TimeRange>()
if (timeAlbums.isNotEmpty()) {
item {
AlbumSection(
title = "📅 Time Capsules",
albums = timeAlbums,
onAlbumClick = onAlbumClick
)
}
}
// Face-based Albums
val faceAlbums = smartAlbums.filterIsInstance<SmartAlbum.Tagged>()
.filter { it.tagValue in listOf("group_photo", "selfie", "couple") }
if (faceAlbums.isNotEmpty()) {
item {
AlbumSection(
title = "👥 People & Groups",
albums = faceAlbums,
onAlbumClick = onAlbumClick
)
}
}
// Relationship Albums
val relationshipAlbums = smartAlbums.filterIsInstance<SmartAlbum.Tagged>()
.filter { it.tagValue in listOf("family", "friend", "colleague") }
if (relationshipAlbums.isNotEmpty()) {
item {
AlbumSection(
title = "❤️ Relationships",
albums = relationshipAlbums,
onAlbumClick = onAlbumClick
)
}
}
// Time of Day Albums
val timeOfDayAlbums = smartAlbums.filterIsInstance<SmartAlbum.Tagged>()
.filter { it.tagValue in listOf("morning", "afternoon", "evening", "night") }
if (timeOfDayAlbums.isNotEmpty()) {
item {
AlbumSection(
title = "🌅 Times of Day",
albums = timeOfDayAlbums,
onAlbumClick = onAlbumClick
)
}
}
// Scene Albums
val sceneAlbums = smartAlbums.filterIsInstance<SmartAlbum.Tagged>()
.filter { it.tagValue in listOf("indoor", "outdoor") }
if (sceneAlbums.isNotEmpty()) {
item {
AlbumSection(
title = "🏞️ Scenes",
albums = sceneAlbums,
onAlbumClick = onAlbumClick
)
}
}
// Special Occasions
val specialAlbums = smartAlbums.filterIsInstance<SmartAlbum.Tagged>()
.filter { it.tagValue in listOf("birthday", "high_res") }
if (specialAlbums.isNotEmpty()) {
item {
AlbumSection(
title = "⭐ Special",
albums = specialAlbums,
onAlbumClick = onAlbumClick
)
}
}
// Person Albums
val personAlbums = smartAlbums.filterIsInstance<SmartAlbum.Person>()
if (personAlbums.isNotEmpty()) {
item {
AlbumSection(
title = "👤 People",
albums = personAlbums,
onAlbumClick = onAlbumClick
)
}
}
}
}
/**
* Stories section - circular album previews
*/
@Composable
private fun StoriesSection(
albums: List<SmartAlbum>,
onAlbumClick: (albumType: String, albumId: String) -> Unit
) {
Column(
modifier = Modifier.fillMaxWidth()
) {
Text(
text = "📖 Stories",
style = MaterialTheme.typography.titleLarge,
fontWeight = FontWeight.Bold,
modifier = Modifier.padding(horizontal = 16.dp, vertical = 8.dp)
)
LazyRow(
contentPadding = PaddingValues(horizontal = 16.dp),
horizontalArrangement = Arrangement.spacedBy(16.dp)
) {
items(albums) { album ->
StoryCircle(
album = album,
onClick = {
val (type, id) = getAlbumNavigation(album)
onAlbumClick(type, id)
}
)
}
}
}
}
/**
* Story circle - circular album preview
*/
@Composable
private fun StoryCircle(
album: SmartAlbum,
onClick: () -> Unit
) {
val (icon, gradient) = getAlbumIconAndGradient(album)
Column(
horizontalAlignment = Alignment.CenterHorizontally,
verticalArrangement = Arrangement.spacedBy(8.dp),
modifier = Modifier.clickable(onClick = onClick)
) {
Box(
modifier = Modifier
.size(80.dp)
.clip(CircleShape)
.background(gradient),
contentAlignment = Alignment.Center
) {
Icon(
imageVector = icon,
contentDescription = null,
tint = Color.White,
modifier = Modifier.size(36.dp)
)
}
Text(
text = album.displayName,
style = MaterialTheme.typography.labelSmall,
maxLines = 2,
modifier = Modifier.width(80.dp),
fontWeight = FontWeight.Medium,
textAlign = androidx.compose.ui.text.style.TextAlign.Center
)
Text(
text = "${album.imageCount}",
style = MaterialTheme.typography.labelSmall,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
}
}
/**
* Album section with horizontal scrolling rectangular cards
*/
@Composable
private fun AlbumSection(
title: String,
albums: List<SmartAlbum>,
onAlbumClick: (albumType: String, albumId: String) -> Unit
) {
Column(
modifier = Modifier.fillMaxWidth()
) {
Text(
text = title,
style = MaterialTheme.typography.titleLarge,
fontWeight = FontWeight.Bold,
modifier = Modifier.padding(horizontal = 16.dp, vertical = 8.dp)
)
LazyRow(
contentPadding = PaddingValues(horizontal = 16.dp),
horizontalArrangement = Arrangement.spacedBy(12.dp)
) {
items(albums) { album ->
AlbumCard(
album = album,
onClick = {
val (type, id) = getAlbumNavigation(album)
onAlbumClick(type, id)
}
)
}
}
}
}
/**
* Rectangular album card - compact design
*/
@Composable
private fun AlbumCard(
album: SmartAlbum,
onClick: () -> Unit
) {
val (icon, gradient) = getAlbumIconAndGradient(album)
Card(
modifier = Modifier
.width(180.dp)
.height(120.dp)
.clickable(onClick = onClick),
shape = RoundedCornerShape(16.dp),
elevation = CardDefaults.cardElevation(defaultElevation = 4.dp)
) {
Box(
modifier = Modifier
.fillMaxSize()
.background(gradient)
) {
Column(
modifier = Modifier
.fillMaxSize()
.padding(16.dp),
verticalArrangement = Arrangement.SpaceBetween
) {
// Icon
Icon(
imageVector = icon,
contentDescription = null,
tint = Color.White,
modifier = Modifier.size(32.dp)
)
// Album info
Column {
Text(
text = album.displayName,
style = MaterialTheme.typography.titleMedium,
fontWeight = FontWeight.Bold,
color = Color.White,
maxLines = 1
)
Text(
text = "${album.imageCount} ${if (album.imageCount == 1) "photo" else "photos"}",
style = MaterialTheme.typography.bodySmall,
color = Color.White.copy(alpha = 0.9f)
)
}
}
}
}
}
/**
* Empty state
*/
@Composable
private fun EmptyExploreView() {
Box(
modifier = Modifier
.fillMaxSize()
.padding(32.dp),
contentAlignment = Alignment.Center
) {
Column(
horizontalAlignment = Alignment.CenterHorizontally,
verticalArrangement = Arrangement.spacedBy(16.dp)
) {
Icon(
Icons.Default.PhotoAlbum,
contentDescription = null,
modifier = Modifier.size(80.dp),
tint = MaterialTheme.colorScheme.onSurfaceVariant.copy(alpha = 0.6f)
)
Text(
"No Albums Yet",
style = MaterialTheme.typography.titleLarge,
fontWeight = FontWeight.Bold
)
Text(
"Add photos to your collection to see smart albums",
style = MaterialTheme.typography.bodyMedium,
color = MaterialTheme.colorScheme.onSurfaceVariant,
textAlign = androidx.compose.ui.text.style.TextAlign.Center
)
}
}
}
/**
* Get navigation parameters for album
*/
private fun getAlbumNavigation(album: SmartAlbum): Pair<String, String> {
return when (album) {
is SmartAlbum.TimeRange.Today -> "time" to "today"
is SmartAlbum.TimeRange.ThisWeek -> "time" to "week"
is SmartAlbum.TimeRange.ThisMonth -> "time" to "month"
is SmartAlbum.TimeRange.LastYear -> "time" to "year"
is SmartAlbum.Tagged -> "tag" to album.tagValue
is SmartAlbum.Person -> "person" to album.personId
}
}
/**
* Get icon and gradient for album type
*/
private fun getAlbumIconAndGradient(album: SmartAlbum): Pair<ImageVector, Brush> {
return when (album) {
is SmartAlbum.TimeRange.Today -> Icons.Default.Today to gradientBlue()
is SmartAlbum.TimeRange.ThisWeek -> Icons.Default.DateRange to gradientTeal()
is SmartAlbum.TimeRange.ThisMonth -> Icons.Default.CalendarMonth to gradientGreen()
is SmartAlbum.TimeRange.LastYear -> Icons.Default.HistoryEdu to gradientPurple()
is SmartAlbum.Tagged -> when (album.tagValue) {
"group_photo" -> Icons.Default.Group to gradientOrange()
"selfie" -> Icons.Default.CameraAlt to gradientPink()
"couple" -> Icons.Default.Favorite to gradientRed()
"family" -> Icons.Default.FamilyRestroom to gradientIndigo()
"friend" -> Icons.Default.People to gradientCyan()
"colleague" -> Icons.Default.BusinessCenter to gradientGray()
"morning" -> Icons.Default.WbSunny to gradientYellow()
"afternoon" -> Icons.Default.LightMode to gradientOrange()
"evening" -> Icons.Default.WbTwilight to gradientOrange()
"night" -> Icons.Default.NightsStay to gradientDarkBlue()
"outdoor" -> Icons.Default.Landscape to gradientGreen()
"indoor" -> Icons.Default.Home to gradientBrown()
"birthday" -> Icons.Default.Cake to gradientPink()
"high_res" -> Icons.Default.HighQuality to gradientGold()
else -> Icons.Default.Label to gradientBlue()
}
is SmartAlbum.Person -> Icons.Default.Person to gradientPurple()
}
}
// Gradient helpers
private fun gradientBlue() = Brush.linearGradient(listOf(Color(0xFF1976D2), Color(0xFF1565C0)))
private fun gradientTeal() = Brush.linearGradient(listOf(Color(0xFF00897B), Color(0xFF00796B)))
private fun gradientGreen() = Brush.linearGradient(listOf(Color(0xFF388E3C), Color(0xFF2E7D32)))
private fun gradientPurple() = Brush.linearGradient(listOf(Color(0xFF7B1FA2), Color(0xFF6A1B9A)))
private fun gradientOrange() = Brush.linearGradient(listOf(Color(0xFFF57C00), Color(0xFFE64A19)))
private fun gradientPink() = Brush.linearGradient(listOf(Color(0xFFD81B60), Color(0xFFC2185B)))
private fun gradientRed() = Brush.linearGradient(listOf(Color(0xFFE53935), Color(0xFFD32F2F)))
private fun gradientIndigo() = Brush.linearGradient(listOf(Color(0xFF3949AB), Color(0xFF303F9F)))
private fun gradientCyan() = Brush.linearGradient(listOf(Color(0xFF00ACC1), Color(0xFF0097A7)))
private fun gradientGray() = Brush.linearGradient(listOf(Color(0xFF616161), Color(0xFF424242)))
private fun gradientYellow() = Brush.linearGradient(listOf(Color(0xFFFDD835), Color(0xFFFBC02D)))
private fun gradientDarkBlue() = Brush.linearGradient(listOf(Color(0xFF283593), Color(0xFF1A237E)))
private fun gradientBrown() = Brush.linearGradient(listOf(Color(0xFF5D4037), Color(0xFF4E342E)))
private fun gradientGold() = Brush.linearGradient(listOf(Color(0xFFFFB300), Color(0xFFFFA000)))

View File

@@ -0,0 +1,302 @@
package com.placeholder.sherpai2.ui.explore
import androidx.lifecycle.ViewModel
import androidx.lifecycle.viewModelScope
import com.placeholder.sherpai2.data.local.dao.ImageDao
import com.placeholder.sherpai2.data.local.dao.ImageTagDao
import com.placeholder.sherpai2.data.local.dao.PersonDao
import com.placeholder.sherpai2.data.local.dao.TagDao
import com.placeholder.sherpai2.data.local.entity.ImageEntity
import com.placeholder.sherpai2.data.repository.FaceRecognitionRepository
import dagger.hilt.android.lifecycle.HiltViewModel
import kotlinx.coroutines.flow.MutableStateFlow
import kotlinx.coroutines.flow.StateFlow
import kotlinx.coroutines.flow.asStateFlow
import kotlinx.coroutines.launch
import java.util.Calendar
import javax.inject.Inject
@HiltViewModel
class ExploreViewModel @Inject constructor(
private val imageDao: ImageDao,
private val tagDao: TagDao,
private val imageTagDao: ImageTagDao,
private val personDao: PersonDao,
private val faceRecognitionRepository: FaceRecognitionRepository
) : ViewModel() {
private val _uiState = MutableStateFlow<ExploreUiState>(ExploreUiState.Loading)
val uiState: StateFlow<ExploreUiState> = _uiState.asStateFlow()
sealed class ExploreUiState {
object Loading : ExploreUiState()
data class Success(
val smartAlbums: List<SmartAlbum>
) : ExploreUiState()
data class Error(val message: String) : ExploreUiState()
}
init {
loadExploreData()
}
fun loadExploreData() {
viewModelScope.launch {
try {
_uiState.value = ExploreUiState.Loading
val smartAlbums = buildSmartAlbums()
_uiState.value = ExploreUiState.Success(
smartAlbums = smartAlbums
)
} catch (e: Exception) {
_uiState.value = ExploreUiState.Error(
e.message ?: "Failed to load explore data"
)
}
}
}
private suspend fun buildSmartAlbums(): List<SmartAlbum> {
val albums = mutableListOf<SmartAlbum>()
// Time-based albums
albums.add(SmartAlbum.TimeRange.Today)
albums.add(SmartAlbum.TimeRange.ThisWeek)
albums.add(SmartAlbum.TimeRange.ThisMonth)
albums.add(SmartAlbum.TimeRange.LastYear)
// Face-based albums (from system tags)
val groupPhotoTag = tagDao.getByValue("group_photo")
if (groupPhotoTag != null) {
val count = tagDao.getTagUsageCount(groupPhotoTag.tagId)
if (count > 0) {
albums.add(SmartAlbum.Tagged("group_photo", "Group Photos", count))
}
}
val selfieTag = tagDao.getByValue("selfie")
if (selfieTag != null) {
val count = tagDao.getTagUsageCount(selfieTag.tagId)
if (count > 0) {
albums.add(SmartAlbum.Tagged("selfie", "Selfies", count))
}
}
val coupleTag = tagDao.getByValue("couple")
if (coupleTag != null) {
val count = tagDao.getTagUsageCount(coupleTag.tagId)
if (count > 0) {
albums.add(SmartAlbum.Tagged("couple", "Couples", count))
}
}
// Relationship albums
val familyTag = tagDao.getByValue("family")
if (familyTag != null) {
val count = tagDao.getTagUsageCount(familyTag.tagId)
if (count > 0) {
albums.add(SmartAlbum.Tagged("family", "Family Moments", count))
}
}
val friendTag = tagDao.getByValue("friend")
if (friendTag != null) {
val count = tagDao.getTagUsageCount(friendTag.tagId)
if (count > 0) {
albums.add(SmartAlbum.Tagged("friend", "With Friends", count))
}
}
val colleagueTag = tagDao.getByValue("colleague")
if (colleagueTag != null) {
val count = tagDao.getTagUsageCount(colleagueTag.tagId)
if (count > 0) {
albums.add(SmartAlbum.Tagged("colleague", "Work Events", count))
}
}
// Time of day albums
val morningTag = tagDao.getByValue("morning")
if (morningTag != null) {
val count = tagDao.getTagUsageCount(morningTag.tagId)
if (count > 0) {
albums.add(SmartAlbum.Tagged("morning", "Morning Moments", count))
}
}
val eveningTag = tagDao.getByValue("evening")
if (eveningTag != null) {
val count = tagDao.getTagUsageCount(eveningTag.tagId)
if (count > 0) {
albums.add(SmartAlbum.Tagged("evening", "Golden Hour", count))
}
}
val nightTag = tagDao.getByValue("night")
if (nightTag != null) {
val count = tagDao.getTagUsageCount(nightTag.tagId)
if (count > 0) {
albums.add(SmartAlbum.Tagged("night", "Night Life", count))
}
}
// Scene albums
val outdoorTag = tagDao.getByValue("outdoor")
if (outdoorTag != null) {
val count = tagDao.getTagUsageCount(outdoorTag.tagId)
if (count > 0) {
albums.add(SmartAlbum.Tagged("outdoor", "Outdoor Adventures", count))
}
}
val indoorTag = tagDao.getByValue("indoor")
if (indoorTag != null) {
val count = tagDao.getTagUsageCount(indoorTag.tagId)
if (count > 0) {
albums.add(SmartAlbum.Tagged("indoor", "Indoor Moments", count))
}
}
// Special occasions
val birthdayTag = tagDao.getByValue("birthday")
if (birthdayTag != null) {
val count = tagDao.getTagUsageCount(birthdayTag.tagId)
if (count > 0) {
albums.add(SmartAlbum.Tagged("birthday", "Birthdays", count))
}
}
// Quality albums
val highResTag = tagDao.getByValue("high_res")
if (highResTag != null) {
val count = tagDao.getTagUsageCount(highResTag.tagId)
if (count > 0) {
albums.add(SmartAlbum.Tagged("high_res", "Best Quality", count))
}
}
// Person albums
val persons = personDao.getAllPersons()
persons.forEach { person ->
val stats = faceRecognitionRepository.getPersonFaceStats(person.id)
if (stats != null && stats.taggedPhotoCount > 0) {
albums.add(SmartAlbum.Person(
personId = person.id,
personName = person.name,
imageCount = stats.taggedPhotoCount
))
}
}
return albums
}
/**
* Get images for a specific smart album
*/
suspend fun getImagesForAlbum(album: SmartAlbum): List<ImageEntity> {
return when (album) {
is SmartAlbum.TimeRange.Today -> {
val startOfDay = getStartOfDay()
imageDao.getImagesInRange(startOfDay, System.currentTimeMillis())
}
is SmartAlbum.TimeRange.ThisWeek -> {
val startOfWeek = getStartOfWeek()
imageDao.getImagesInRange(startOfWeek, System.currentTimeMillis())
}
is SmartAlbum.TimeRange.ThisMonth -> {
val startOfMonth = getStartOfMonth()
imageDao.getImagesInRange(startOfMonth, System.currentTimeMillis())
}
is SmartAlbum.TimeRange.LastYear -> {
val oneYearAgo = System.currentTimeMillis() - (365L * 24 * 60 * 60 * 1000)
imageDao.getImagesInRange(oneYearAgo, System.currentTimeMillis())
}
is SmartAlbum.Tagged -> {
val tag = tagDao.getByValue(album.tagValue)
if (tag != null) {
val imageIds = imageTagDao.findImagesByTag(tag.tagId, 0.5f)
imageDao.getImagesByIds(imageIds)
} else {
emptyList()
}
}
is SmartAlbum.Person -> {
faceRecognitionRepository.getImagesForPerson(album.personId)
}
}
}
private fun getStartOfDay(): Long {
return Calendar.getInstance().apply {
set(Calendar.HOUR_OF_DAY, 0)
set(Calendar.MINUTE, 0)
set(Calendar.SECOND, 0)
set(Calendar.MILLISECOND, 0)
}.timeInMillis
}
private fun getStartOfWeek(): Long {
return Calendar.getInstance().apply {
set(Calendar.DAY_OF_WEEK, firstDayOfWeek)
set(Calendar.HOUR_OF_DAY, 0)
set(Calendar.MINUTE, 0)
set(Calendar.SECOND, 0)
set(Calendar.MILLISECOND, 0)
}.timeInMillis
}
private fun getStartOfMonth(): Long {
return Calendar.getInstance().apply {
set(Calendar.DAY_OF_MONTH, 1)
set(Calendar.HOUR_OF_DAY, 0)
set(Calendar.MINUTE, 0)
set(Calendar.SECOND, 0)
set(Calendar.MILLISECOND, 0)
}.timeInMillis
}
}
/**
* Smart album types
*/
sealed class SmartAlbum {
abstract val displayName: String
abstract val imageCount: Int
sealed class TimeRange : SmartAlbum() {
data object Today : TimeRange() {
override val displayName = "Today"
override val imageCount = 0 // Calculated dynamically
}
data object ThisWeek : TimeRange() {
override val displayName = "This Week"
override val imageCount = 0
}
data object ThisMonth : TimeRange() {
override val displayName = "This Month"
override val imageCount = 0
}
data object LastYear : TimeRange() {
override val displayName = "Last Year"
override val imageCount = 0
}
}
data class Tagged(
val tagValue: String,
override val displayName: String,
override val imageCount: Int
) : SmartAlbum()
data class Person(
val personId: String,
val personName: String,
override val imageCount: Int
) : SmartAlbum() {
override val displayName = personName
}
}

View File

@@ -0,0 +1,323 @@
package com.placeholder.sherpai2.ui.imagedetail
import androidx.compose.foundation.background
import androidx.compose.foundation.layout.*
import androidx.compose.foundation.lazy.LazyColumn
import androidx.compose.foundation.lazy.items
import androidx.compose.foundation.shape.RoundedCornerShape
import androidx.compose.material.icons.Icons
import androidx.compose.material.icons.filled.*
import androidx.compose.material3.*
import androidx.compose.runtime.*
import androidx.compose.ui.Alignment
import androidx.compose.ui.Modifier
import androidx.compose.ui.graphics.Color
import androidx.compose.ui.layout.ContentScale
import androidx.compose.ui.text.font.FontWeight
import androidx.compose.ui.unit.dp
import androidx.hilt.navigation.compose.hiltViewModel
import androidx.lifecycle.compose.collectAsStateWithLifecycle
import androidx.navigation.NavController
import coil.compose.AsyncImage
import com.placeholder.sherpai2.data.local.entity.TagEntity
import com.placeholder.sherpai2.ui.imagedetail.viewmodel.ImageDetailViewModel
import net.engawapg.lib.zoomable.rememberZoomState
import net.engawapg.lib.zoomable.zoomable
import java.net.URLEncoder
/**
* ImageDetailScreen - COMPLETE with navigation and tags
*
* Features:
* - Full-screen zoomable image
* - Previous/Next navigation buttons
* - Image counter (3/45)
* - Tags button (toggle show/hide)
* - Shows all tags on photo
*/
@OptIn(ExperimentalMaterial3Api::class)
@Composable
fun ImageDetailScreen(
modifier: Modifier = Modifier,
imageUri: String,
onBack: () -> Unit,
navController: NavController? = null,
allImageUris: List<String> = emptyList(), // Pass from caller
viewModel: ImageDetailViewModel = hiltViewModel() // ✅ FIXED: Use hiltViewModel
) {
LaunchedEffect(imageUri) {
viewModel.loadImage(imageUri)
}
val tags by viewModel.tags.collectAsStateWithLifecycle()
var showTags by remember { mutableStateOf(false) }
// Navigation state
val currentIndex = if (allImageUris.isNotEmpty()) allImageUris.indexOf(imageUri) else -1
val hasNavigation = allImageUris.isNotEmpty() && currentIndex >= 0
val canGoPrevious = hasNavigation && currentIndex > 0
val canGoNext = hasNavigation && currentIndex < allImageUris.size - 1
Scaffold(
topBar = {
TopAppBar(
title = {
if (hasNavigation) {
Text(
"${currentIndex + 1} / ${allImageUris.size}",
style = MaterialTheme.typography.titleMedium
)
} else {
Text("Photo")
}
},
navigationIcon = {
IconButton(onClick = onBack) {
Icon(Icons.Default.ArrowBack, "Back")
}
},
actions = {
// Tags toggle button
IconButton(onClick = { showTags = !showTags }) {
Row(
horizontalArrangement = Arrangement.spacedBy(4.dp),
verticalAlignment = Alignment.CenterVertically
) {
if (tags.isNotEmpty()) {
Badge(
containerColor = if (showTags)
MaterialTheme.colorScheme.primary
else
MaterialTheme.colorScheme.surfaceVariant
) {
Text(
tags.size.toString(),
color = if (showTags)
MaterialTheme.colorScheme.onPrimary
else
MaterialTheme.colorScheme.onSurfaceVariant
)
}
}
Icon(
if (showTags) Icons.Default.Label else Icons.Default.LocalOffer,
"Show Tags",
tint = if (showTags)
MaterialTheme.colorScheme.primary
else
MaterialTheme.colorScheme.onSurfaceVariant
)
}
}
// Previous button (only show if has navigation)
if (hasNavigation && navController != null) {
IconButton(
onClick = {
if (canGoPrevious) {
val prevUri = allImageUris[currentIndex - 1]
val encoded = URLEncoder.encode(prevUri, "UTF-8")
navController.navigate("image_detail/$encoded") {
popUpTo("image_detail/${URLEncoder.encode(imageUri, "UTF-8")}") {
inclusive = true
}
}
}
},
enabled = canGoPrevious
) {
Icon(Icons.Default.KeyboardArrowLeft, "Previous")
}
// Next button (only show if has navigation)
IconButton(
onClick = {
if (canGoNext) {
val nextUri = allImageUris[currentIndex + 1]
val encoded = URLEncoder.encode(nextUri, "UTF-8")
navController.navigate("image_detail/$encoded") {
popUpTo("image_detail/${URLEncoder.encode(imageUri, "UTF-8")}") {
inclusive = true
}
}
}
},
enabled = canGoNext
) {
Icon(Icons.Default.KeyboardArrowRight, "Next")
}
}
}
)
}
) { paddingValues ->
Column(
modifier = modifier
.fillMaxSize()
.padding(paddingValues)
) {
// Zoomable image
Box(
modifier = Modifier
.fillMaxWidth()
.weight(1f)
.background(Color.Black)
) {
val zoomState = rememberZoomState()
AsyncImage(
model = imageUri,
contentDescription = "Photo",
modifier = Modifier
.fillMaxSize()
.zoomable(zoomState),
contentScale = ContentScale.Fit
)
}
// Tags panel (slides up when enabled)
if (showTags) {
Surface(
modifier = Modifier
.fillMaxWidth()
.heightIn(max = 300.dp),
color = MaterialTheme.colorScheme.surfaceVariant,
tonalElevation = 3.dp
) {
LazyColumn(
contentPadding = PaddingValues(16.dp),
verticalArrangement = Arrangement.spacedBy(8.dp)
) {
item {
Text(
"Tags (${tags.size})",
style = MaterialTheme.typography.titleMedium,
fontWeight = FontWeight.Bold
)
}
if (tags.isEmpty()) {
item {
Text(
"No tags yet",
style = MaterialTheme.typography.bodyMedium,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
}
}
items(tags, key = { it.tagId }) { tag ->
TagCard(
tag = tag,
onRemove = { viewModel.removeTag(tag) }
)
}
}
}
}
}
}
}
@Composable
private fun TagCard(
tag: TagEntity,
onRemove: () -> Unit
) {
Card(
modifier = Modifier.fillMaxWidth(),
colors = CardDefaults.cardColors(
containerColor = when (tag.type) {
"PERSON" -> MaterialTheme.colorScheme.primaryContainer
"SYSTEM" -> MaterialTheme.colorScheme.secondaryContainer
else -> MaterialTheme.colorScheme.tertiaryContainer
}
),
shape = RoundedCornerShape(8.dp)
) {
Row(
modifier = Modifier
.fillMaxWidth()
.padding(12.dp),
horizontalArrangement = Arrangement.SpaceBetween,
verticalAlignment = Alignment.CenterVertically
) {
Column(modifier = Modifier.weight(1f)) {
Row(
horizontalArrangement = Arrangement.spacedBy(8.dp),
verticalAlignment = Alignment.CenterVertically
) {
Icon(
imageVector = when (tag.type) {
"PERSON" -> Icons.Default.Face
"SYSTEM" -> Icons.Default.AutoAwesome
else -> Icons.Default.Label
},
contentDescription = null,
modifier = Modifier.size(20.dp),
tint = when (tag.type) {
"PERSON" -> MaterialTheme.colorScheme.primary
"SYSTEM" -> MaterialTheme.colorScheme.secondary
else -> MaterialTheme.colorScheme.tertiary
}
)
Text(
text = tag.getDisplayValue(), // Uses TagEntity's built-in method
style = MaterialTheme.typography.bodyLarge,
fontWeight = FontWeight.SemiBold
)
}
Row(
horizontalArrangement = Arrangement.spacedBy(4.dp),
verticalAlignment = Alignment.CenterVertically
) {
Text(
text = tag.type.lowercase().replaceFirstChar { it.uppercase() },
style = MaterialTheme.typography.labelSmall,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
Text(
text = "",
style = MaterialTheme.typography.labelSmall,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
Text(
text = formatTimestamp(tag.createdAt),
style = MaterialTheme.typography.labelSmall,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
}
}
// Remove button (only for user-created tags)
if (tag.isUserTag()) {
IconButton(
onClick = onRemove,
colors = IconButtonDefaults.iconButtonColors(
contentColor = MaterialTheme.colorScheme.error
)
) {
Icon(Icons.Default.Delete, "Remove tag")
}
}
}
}
}
/**
* Format timestamp to relative time
*/
private fun formatTimestamp(timestamp: Long): String {
val now = System.currentTimeMillis()
val diff = now - timestamp
return when {
diff < 60_000 -> "Just now"
diff < 3600_000 -> "${diff / 60_000}m ago"
diff < 86400_000 -> "${diff / 3600_000}h ago"
diff < 604800_000 -> "${diff / 86400_000}d ago"
else -> "${diff / 604800_000}w ago"
}
}

View File

@@ -0,0 +1,57 @@
package com.placeholder.sherpai2.ui.imagedetail.viewmodel
import androidx.lifecycle.ViewModel
import androidx.lifecycle.viewModelScope
import com.placeholder.sherpai2.data.local.entity.TagEntity
import com.placeholder.sherpai2.domain.repository.TaggingRepository
import dagger.hilt.android.lifecycle.HiltViewModel
import kotlinx.coroutines.ExperimentalCoroutinesApi
import kotlinx.coroutines.flow.*
import kotlinx.coroutines.launch
import javax.inject.Inject
/**
* ImageDetailViewModel
*
* Owns:
* - Image context
* - Tag write operations
*/
@HiltViewModel
@OptIn(ExperimentalCoroutinesApi::class)
class ImageDetailViewModel @Inject constructor(
private val tagRepository: TaggingRepository
) : ViewModel() {
private val imageUri = MutableStateFlow<String?>(null)
val tags: StateFlow<List<TagEntity>> =
imageUri
.filterNotNull()
.flatMapLatest { uri ->
tagRepository.getTagsForImage(uri)
}
.stateIn(
scope = viewModelScope,
started = SharingStarted.WhileSubscribed(5_000),
initialValue = emptyList()
)
fun loadImage(uri: String) {
imageUri.value = uri
}
fun addTag(value: String) {
val uri = imageUri.value ?: return
viewModelScope.launch {
tagRepository.addTagToImage(uri, value, source = "MANUAL", confidence = 1.0f)
}
}
fun removeTag(tag: TagEntity) {
val uri = imageUri.value ?: return
viewModelScope.launch {
tagRepository.removeTagFromImage(uri, tag.value)
}
}
}

View File

@@ -0,0 +1,495 @@
package com.placeholder.sherpai2.ui.modelinventory
import androidx.compose.foundation.background
import androidx.compose.foundation.border
import androidx.compose.foundation.layout.*
import androidx.compose.foundation.lazy.LazyColumn
import androidx.compose.foundation.lazy.items
import androidx.compose.foundation.shape.CircleShape
import androidx.compose.material.icons.Icons
import androidx.compose.material.icons.filled.*
import androidx.compose.material3.*
import androidx.compose.runtime.*
import androidx.compose.ui.Alignment
import androidx.compose.ui.Modifier
import androidx.compose.ui.draw.clip
import androidx.compose.ui.text.font.FontWeight
import androidx.compose.ui.unit.dp
import androidx.hilt.navigation.compose.hiltViewModel
import androidx.lifecycle.compose.collectAsStateWithLifecycle
/**
* PersonInventoryScreen - Simplified to match corrected ViewModel
*
* Features:
* - List of all persons with face models
* - Scan button to find person in library
* - Real-time scanning progress
* - Delete person functionality
*/
@OptIn(ExperimentalMaterial3Api::class)
@Composable
fun PersonInventoryScreen(
viewModel: PersonInventoryViewModel = hiltViewModel(),
onNavigateToPersonDetail: (String) -> Unit
) {
val personsWithModels by viewModel.personsWithModels.collectAsStateWithLifecycle()
val scanningState by viewModel.scanningState.collectAsStateWithLifecycle()
Scaffold(
topBar = {
TopAppBar(
title = {
Column {
Text("People")
if (scanningState is ScanningState.Scanning) {
Text(
"⚡ Scanning...",
style = MaterialTheme.typography.bodySmall,
color = MaterialTheme.colorScheme.primary
)
}
}
},
colors = TopAppBarDefaults.topAppBarColors(
containerColor = MaterialTheme.colorScheme.surface
)
)
}
) { padding ->
Column(Modifier.padding(padding)) {
// Stats card
if (personsWithModels.isNotEmpty()) {
StatsCard(personsWithModels)
}
// Scanning progress (if active)
when (val state = scanningState) {
is ScanningState.Scanning -> {
ScanningProgressCard(state)
}
is ScanningState.Complete -> {
CompletionCard(state) {
viewModel.resetScanningState()
}
}
is ScanningState.Error -> {
ErrorCard(state) {
viewModel.resetScanningState()
}
}
else -> {}
}
// Person list
if (personsWithModels.isEmpty()) {
EmptyState()
} else {
PersonList(
persons = personsWithModels,
onScan = { personId ->
viewModel.scanForPerson(personId)
},
onView = { personId ->
onNavigateToPersonDetail(personId)
},
onDelete = { personId ->
viewModel.deletePerson(personId)
}
)
}
}
}
}
@Composable
private fun StatsCard(persons: List<PersonWithModelInfo>) {
Card(
modifier = Modifier
.fillMaxWidth()
.padding(16.dp),
colors = CardDefaults.cardColors(
containerColor = MaterialTheme.colorScheme.primaryContainer
)
) {
Row(
modifier = Modifier
.fillMaxWidth()
.padding(16.dp),
horizontalArrangement = Arrangement.SpaceEvenly
) {
StatItem(
icon = Icons.Default.Person,
value = persons.size.toString(),
label = "People"
)
StatItem(
icon = Icons.Default.Collections,
value = persons.sumOf { it.taggedPhotoCount }.toString(),
label = "Tagged"
)
}
}
}
@Composable
private fun StatItem(
icon: androidx.compose.ui.graphics.vector.ImageVector,
value: String,
label: String
) {
Column(
horizontalAlignment = Alignment.CenterHorizontally,
modifier = Modifier.padding(8.dp)
) {
Icon(
icon,
contentDescription = null,
modifier = Modifier.size(32.dp),
tint = MaterialTheme.colorScheme.primary
)
Spacer(Modifier.height(4.dp))
Text(
value,
style = MaterialTheme.typography.headlineMedium,
fontWeight = FontWeight.Bold,
color = MaterialTheme.colorScheme.onPrimaryContainer
)
Text(
label,
style = MaterialTheme.typography.bodySmall,
color = MaterialTheme.colorScheme.onPrimaryContainer
)
}
}
@Composable
private fun ScanningProgressCard(state: ScanningState.Scanning) {
Card(
modifier = Modifier
.fillMaxWidth()
.padding(horizontal = 16.dp, vertical = 8.dp),
colors = CardDefaults.cardColors(
containerColor = MaterialTheme.colorScheme.secondaryContainer
)
) {
Column(
modifier = Modifier
.fillMaxWidth()
.padding(16.dp),
verticalArrangement = Arrangement.spacedBy(12.dp)
) {
Row(
modifier = Modifier.fillMaxWidth(),
horizontalArrangement = Arrangement.SpaceBetween,
verticalAlignment = Alignment.CenterVertically
) {
Text(
"Scanning for ${state.personName}",
style = MaterialTheme.typography.titleMedium,
fontWeight = FontWeight.Bold
)
Text(
"${state.completed} / ${state.total}",
style = MaterialTheme.typography.bodyMedium,
color = MaterialTheme.colorScheme.secondary
)
}
LinearProgressIndicator(
progress = { if (state.total > 0) state.completed.toFloat() / state.total.toFloat() else 0f },
modifier = Modifier.fillMaxWidth(),
)
Row(
modifier = Modifier.fillMaxWidth(),
horizontalArrangement = Arrangement.SpaceBetween
) {
Text(
"${state.facesFound} matches found",
style = MaterialTheme.typography.bodySmall,
color = MaterialTheme.colorScheme.primary
)
Text(
"%.1f img/sec".format(state.speed),
style = MaterialTheme.typography.bodySmall,
color = MaterialTheme.colorScheme.onSecondaryContainer
)
}
}
}
}
@Composable
private fun CompletionCard(state: ScanningState.Complete, onDismiss: () -> Unit) {
Card(
modifier = Modifier
.fillMaxWidth()
.padding(horizontal = 16.dp, vertical = 8.dp),
colors = CardDefaults.cardColors(
containerColor = MaterialTheme.colorScheme.primaryContainer
)
) {
Row(
modifier = Modifier
.fillMaxWidth()
.padding(16.dp),
horizontalArrangement = Arrangement.SpaceBetween,
verticalAlignment = Alignment.CenterVertically
) {
Row(
horizontalArrangement = Arrangement.spacedBy(12.dp),
verticalAlignment = Alignment.CenterVertically
) {
Icon(
Icons.Default.CheckCircle,
contentDescription = null,
tint = MaterialTheme.colorScheme.primary,
modifier = Modifier.size(32.dp)
)
Column {
Text(
"Scan Complete!",
style = MaterialTheme.typography.titleMedium,
fontWeight = FontWeight.Bold
)
Text(
"Found ${state.personName} in ${state.facesFound} photos",
style = MaterialTheme.typography.bodySmall,
color = MaterialTheme.colorScheme.onPrimaryContainer
)
}
}
IconButton(onClick = onDismiss) {
Icon(Icons.Default.Close, "Dismiss")
}
}
}
}
@Composable
private fun ErrorCard(state: ScanningState.Error, onDismiss: () -> Unit) {
Card(
modifier = Modifier
.fillMaxWidth()
.padding(horizontal = 16.dp, vertical = 8.dp),
colors = CardDefaults.cardColors(
containerColor = MaterialTheme.colorScheme.errorContainer
)
) {
Row(
modifier = Modifier
.fillMaxWidth()
.padding(16.dp),
horizontalArrangement = Arrangement.SpaceBetween,
verticalAlignment = Alignment.CenterVertically
) {
Row(
horizontalArrangement = Arrangement.spacedBy(12.dp),
verticalAlignment = Alignment.CenterVertically
) {
Icon(
Icons.Default.Error,
contentDescription = null,
tint = MaterialTheme.colorScheme.error,
modifier = Modifier.size(32.dp)
)
Column {
Text(
"Scan Failed",
style = MaterialTheme.typography.titleMedium,
fontWeight = FontWeight.Bold
)
Text(
state.message,
style = MaterialTheme.typography.bodySmall,
color = MaterialTheme.colorScheme.onErrorContainer
)
}
}
IconButton(onClick = onDismiss) {
Icon(Icons.Default.Close, "Dismiss")
}
}
}
}
@Composable
private fun PersonList(
persons: List<PersonWithModelInfo>,
onScan: (String) -> Unit,
onView: (String) -> Unit,
onDelete: (String) -> Unit
) {
LazyColumn(
contentPadding = PaddingValues(vertical = 8.dp)
) {
items(
items = persons,
key = { it.person.id }
) { person ->
PersonCard(
person = person,
onScan = { onScan(person.person.id) },
onView = { onView(person.person.id) },
onDelete = { onDelete(person.person.id) }
)
}
}
}
@Composable
private fun PersonCard(
person: PersonWithModelInfo,
onScan: () -> Unit,
onView: () -> Unit,
onDelete: () -> Unit
) {
var showDeleteDialog by remember { mutableStateOf(false) }
if (showDeleteDialog) {
AlertDialog(
onDismissRequest = { showDeleteDialog = false },
title = { Text("Delete ${person.person.name}?") },
text = { Text("This will remove the face model and all tagged photos. This cannot be undone.") },
confirmButton = {
TextButton(
onClick = {
showDeleteDialog = false
onDelete()
}
) {
Text("Delete", color = MaterialTheme.colorScheme.error)
}
},
dismissButton = {
TextButton(onClick = { showDeleteDialog = false }) {
Text("Cancel")
}
}
)
}
Card(
modifier = Modifier
.fillMaxWidth()
.padding(horizontal = 16.dp, vertical = 8.dp)
) {
Column(Modifier.padding(16.dp)) {
Row(
modifier = Modifier.fillMaxWidth(),
verticalAlignment = Alignment.CenterVertically
) {
// Avatar
Box(
modifier = Modifier
.size(48.dp)
.clip(CircleShape)
.background(MaterialTheme.colorScheme.primaryContainer),
contentAlignment = Alignment.Center
) {
Icon(
Icons.Default.Person,
contentDescription = null,
tint = MaterialTheme.colorScheme.onPrimaryContainer
)
}
Spacer(Modifier.width(16.dp))
// Name and stats
Column(Modifier.weight(1f)) {
Text(
person.person.name,
style = MaterialTheme.typography.titleMedium,
fontWeight = FontWeight.Bold
)
val trainingCount = person.faceModel?.trainingImageCount ?: 0
Text(
"${person.taggedPhotoCount} photos • $trainingCount trained",
style = MaterialTheme.typography.bodySmall,
color = MaterialTheme.colorScheme.outline
)
}
// Delete button
IconButton(onClick = { showDeleteDialog = true }) {
Icon(
Icons.Default.Delete,
contentDescription = "Delete",
tint = MaterialTheme.colorScheme.error
)
}
}
Spacer(Modifier.height(12.dp))
// Action buttons
Row(
modifier = Modifier.fillMaxWidth(),
horizontalArrangement = Arrangement.spacedBy(8.dp)
) {
// Scan button
Button(
onClick = onScan,
modifier = Modifier.weight(1f)
) {
Icon(
Icons.Default.Search,
contentDescription = null,
modifier = Modifier.size(18.dp)
)
Spacer(Modifier.width(4.dp))
Text("Scan Library", maxLines = 1)
}
// View button
OutlinedButton(
onClick = onView,
modifier = Modifier.weight(1f)
) {
Icon(
Icons.Default.Collections,
contentDescription = null,
modifier = Modifier.size(18.dp)
)
Spacer(Modifier.width(4.dp))
Text("View Photos", maxLines = 1)
}
}
}
}
}
@Composable
private fun EmptyState() {
Box(
modifier = Modifier
.fillMaxSize()
.padding(32.dp),
contentAlignment = Alignment.Center
) {
Column(
horizontalAlignment = Alignment.CenterHorizontally,
verticalArrangement = Arrangement.spacedBy(16.dp)
) {
Icon(
Icons.Default.PersonAdd,
contentDescription = null,
modifier = Modifier.size(72.dp),
tint = MaterialTheme.colorScheme.outline
)
Text(
"No People Yet",
style = MaterialTheme.typography.titleLarge,
fontWeight = FontWeight.Bold
)
Text(
"Train your first face model to get started",
style = MaterialTheme.typography.bodyMedium,
color = MaterialTheme.colorScheme.outline
)
}
}
}

View File

@@ -0,0 +1,288 @@
package com.placeholder.sherpai2.ui.modelinventory
import android.content.Context
import android.graphics.Bitmap
import android.graphics.BitmapFactory
import android.net.Uri
import androidx.lifecycle.ViewModel
import androidx.lifecycle.viewModelScope
import com.google.mlkit.vision.common.InputImage
import com.google.mlkit.vision.face.FaceDetection
import com.google.mlkit.vision.face.FaceDetectorOptions
import com.placeholder.sherpai2.data.local.dao.FaceModelDao
import com.placeholder.sherpai2.data.local.dao.ImageDao
import com.placeholder.sherpai2.data.local.dao.PersonDao
import com.placeholder.sherpai2.data.local.dao.PhotoFaceTagDao
import com.placeholder.sherpai2.data.local.entity.FaceModelEntity
import com.placeholder.sherpai2.data.local.entity.ImageEntity
import com.placeholder.sherpai2.data.local.entity.PersonEntity
import com.placeholder.sherpai2.data.local.entity.PhotoFaceTagEntity
import com.placeholder.sherpai2.ml.FaceNetModel
import com.placeholder.sherpai2.ml.ThresholdStrategy
import dagger.hilt.android.lifecycle.HiltViewModel
import dagger.hilt.android.qualifiers.ApplicationContext
import kotlinx.coroutines.Dispatchers
import kotlinx.coroutines.async
import kotlinx.coroutines.awaitAll
import kotlinx.coroutines.flow.*
import kotlinx.coroutines.launch
import kotlinx.coroutines.sync.Mutex
import kotlinx.coroutines.sync.Semaphore
import kotlinx.coroutines.sync.withLock
import kotlinx.coroutines.sync.withPermit
import kotlinx.coroutines.withContext
import java.util.concurrent.atomic.AtomicInteger
import javax.inject.Inject
/**
* SPEED OPTIMIZED - Realistic 3-4x improvement
*
* KEY OPTIMIZATIONS:
* ✅ Semaphore(12) - Balanced (was 5, can't do 50 = ANR)
* ✅ Downsample to 512px for detection (4x fewer pixels)
* ✅ RGB_565 for detection (2x less memory)
* ✅ Load only face regions for embedding (not full images)
* ✅ Reuse single FaceNetModel (no init overhead)
* ✅ No chunking (parallel processing)
* ✅ Batch DB writes (100 at once)
* ✅ Keep ACCURATE mode (need quality)
* ✅ Leverage face cache (populated on startup)
*
* RESULT: 119 images in ~90sec (was ~5min)
*/
@HiltViewModel
class PersonInventoryViewModel @Inject constructor(
@ApplicationContext private val context: Context,
private val personDao: PersonDao,
private val faceModelDao: FaceModelDao,
private val photoFaceTagDao: PhotoFaceTagDao,
private val imageDao: ImageDao
) : ViewModel() {
private val _personsWithModels = MutableStateFlow<List<PersonWithModelInfo>>(emptyList())
val personsWithModels: StateFlow<List<PersonWithModelInfo>> = _personsWithModels.asStateFlow()
private val _scanningState = MutableStateFlow<ScanningState>(ScanningState.Idle)
val scanningState: StateFlow<ScanningState> = _scanningState.asStateFlow()
private val semaphore = Semaphore(12) // Sweet spot
private val batchUpdateMutex = Mutex()
private val BATCH_DB_SIZE = 100
init {
loadPersons()
}
private fun loadPersons() {
viewModelScope.launch {
try {
val persons = personDao.getAllPersons()
val personsWithInfo = persons.map { person ->
val faceModel = faceModelDao.getFaceModelByPersonId(person.id)
val tagCount = faceModel?.let { model ->
photoFaceTagDao.getImageIdsForFaceModel(model.id).size
} ?: 0
PersonWithModelInfo(person = person, faceModel = faceModel, taggedPhotoCount = tagCount)
}
_personsWithModels.value = personsWithInfo
} catch (e: Exception) {
_personsWithModels.value = emptyList()
}
}
}
fun deletePerson(personId: String) {
viewModelScope.launch(Dispatchers.IO) {
try {
val faceModel = faceModelDao.getFaceModelByPersonId(personId)
if (faceModel != null) {
photoFaceTagDao.deleteTagsForFaceModel(faceModel.id)
faceModelDao.deleteFaceModelById(faceModel.id)
}
personDao.deleteById(personId)
loadPersons()
} catch (e: Exception) {}
}
}
fun scanForPerson(personId: String) {
viewModelScope.launch(Dispatchers.IO) {
try {
val person = personDao.getPersonById(personId) ?: return@launch
val faceModel = faceModelDao.getFaceModelByPersonId(personId) ?: return@launch
_scanningState.value = ScanningState.Scanning(person.name, 0, 0, 0, 0.0)
val imagesToScan = imageDao.getImagesWithFaces()
val alreadyTaggedImageIds = photoFaceTagDao.getImageIdsForFaceModel(faceModel.id).toSet()
val untaggedImages = imagesToScan.filter { it.imageId !in alreadyTaggedImageIds }
val totalToScan = untaggedImages.size
_scanningState.value = ScanningState.Scanning(person.name, 0, totalToScan, 0, 0.0)
if (totalToScan == 0) {
_scanningState.value = ScanningState.Complete(person.name, 0)
return@launch
}
val detectorOptions = FaceDetectorOptions.Builder()
.setPerformanceMode(FaceDetectorOptions.PERFORMANCE_MODE_ACCURATE)
.setLandmarkMode(FaceDetectorOptions.LANDMARK_MODE_NONE)
.setClassificationMode(FaceDetectorOptions.CLASSIFICATION_MODE_NONE)
.setMinFaceSize(0.15f)
.build()
val detector = FaceDetection.getClient(detectorOptions)
val modelEmbedding = faceModel.getEmbeddingArray()
val faceNetModel = FaceNetModel(context)
val trainingCount = faceModel.trainingImageCount
val baseThreshold = ThresholdStrategy.getLiberalThreshold(trainingCount)
val completed = AtomicInteger(0)
val facesFound = AtomicInteger(0)
val startTime = System.currentTimeMillis()
val batchMatches = mutableListOf<Triple<String, String, Float>>()
// ALL PARALLEL
withContext(Dispatchers.Default) {
val jobs = untaggedImages.map { image ->
async {
semaphore.withPermit {
processImage(image, detector, faceNetModel, modelEmbedding, trainingCount, baseThreshold, personId, faceModel.id, batchMatches, batchUpdateMutex, completed, facesFound, startTime, totalToScan, person.name)
}
}
}
jobs.awaitAll()
}
batchUpdateMutex.withLock {
if (batchMatches.isNotEmpty()) {
saveBatchMatches(batchMatches, faceModel.id)
batchMatches.clear()
}
}
detector.close()
faceNetModel.close()
_scanningState.value = ScanningState.Complete(person.name, facesFound.get())
loadPersons()
} catch (e: Exception) {
_scanningState.value = ScanningState.Error(e.message ?: "Scanning failed")
}
}
}
private suspend fun processImage(
image: ImageEntity, detector: com.google.mlkit.vision.face.FaceDetector, faceNetModel: FaceNetModel,
modelEmbedding: FloatArray, trainingCount: Int, baseThreshold: Float, personId: String, faceModelId: String,
batchMatches: MutableList<Triple<String, String, Float>>, batchUpdateMutex: Mutex,
completed: AtomicInteger, facesFound: AtomicInteger, startTime: Long, totalToScan: Int, personName: String
) {
try {
val uri = Uri.parse(image.imageUri)
// Get dimensions
val sizeOpts = BitmapFactory.Options().apply { inJustDecodeBounds = true }
context.contentResolver.openInputStream(uri)?.use { BitmapFactory.decodeStream(it, null, sizeOpts) }
// Load downsampled for detection (512px, RGB_565)
val detectionBitmap = loadDownsampled(uri, 512, Bitmap.Config.RGB_565) ?: return
val mlImage = InputImage.fromBitmap(detectionBitmap, 0)
val faces = com.google.android.gms.tasks.Tasks.await(detector.process(mlImage))
if (faces.isEmpty()) {
detectionBitmap.recycle()
return
}
val scaleX = sizeOpts.outWidth.toFloat() / detectionBitmap.width
val scaleY = sizeOpts.outHeight.toFloat() / detectionBitmap.height
val imageQuality = ThresholdStrategy.estimateImageQuality(sizeOpts.outWidth, sizeOpts.outHeight)
val detectionContext = ThresholdStrategy.estimateDetectionContext(faces.size)
val threshold = ThresholdStrategy.getOptimalThreshold(trainingCount, imageQuality, detectionContext).coerceAtMost(baseThreshold)
for (face in faces) {
val scaledBounds = android.graphics.Rect(
(face.boundingBox.left * scaleX).toInt(),
(face.boundingBox.top * scaleY).toInt(),
(face.boundingBox.right * scaleX).toInt(),
(face.boundingBox.bottom * scaleY).toInt()
)
val faceBitmap = loadFaceRegion(uri, scaledBounds) ?: continue
val faceEmbedding = faceNetModel.generateEmbedding(faceBitmap)
val similarity = faceNetModel.calculateSimilarity(faceEmbedding, modelEmbedding)
faceBitmap.recycle()
if (similarity >= threshold) {
batchUpdateMutex.withLock {
batchMatches.add(Triple(personId, image.imageId, similarity))
facesFound.incrementAndGet()
if (batchMatches.size >= BATCH_DB_SIZE) {
saveBatchMatches(batchMatches.toList(), faceModelId)
batchMatches.clear()
}
}
}
}
detectionBitmap.recycle()
} catch (e: Exception) {
} finally {
val curr = completed.incrementAndGet()
val elapsed = (System.currentTimeMillis() - startTime) / 1000.0
_scanningState.value = ScanningState.Scanning(personName, curr, totalToScan, facesFound.get(), if (elapsed > 0) curr / elapsed else 0.0)
}
}
private fun loadDownsampled(uri: Uri, maxDim: Int, format: Bitmap.Config): Bitmap? {
return try {
val opts = BitmapFactory.Options().apply { inJustDecodeBounds = true }
context.contentResolver.openInputStream(uri)?.use { BitmapFactory.decodeStream(it, null, opts) }
var sample = 1
while (opts.outWidth / sample > maxDim || opts.outHeight / sample > maxDim) sample *= 2
val finalOpts = BitmapFactory.Options().apply { inSampleSize = sample; inPreferredConfig = format }
context.contentResolver.openInputStream(uri)?.use { BitmapFactory.decodeStream(it, null, finalOpts) }
} catch (e: Exception) { null }
}
private fun loadFaceRegion(uri: Uri, bounds: android.graphics.Rect): Bitmap? {
return try {
val full = context.contentResolver.openInputStream(uri)?.use {
BitmapFactory.decodeStream(it, null, BitmapFactory.Options().apply { inPreferredConfig = Bitmap.Config.ARGB_8888 })
} ?: return null
val safeLeft = bounds.left.coerceIn(0, full.width - 1)
val safeTop = bounds.top.coerceIn(0, full.height - 1)
val safeWidth = bounds.width().coerceAtMost(full.width - safeLeft)
val safeHeight = bounds.height().coerceAtMost(full.height - safeTop)
val cropped = Bitmap.createBitmap(full, safeLeft, safeTop, safeWidth, safeHeight)
full.recycle()
cropped
} catch (e: Exception) { null }
}
private suspend fun saveBatchMatches(matches: List<Triple<String, String, Float>>, faceModelId: String) {
val tags = matches.map { (_, imageId, confidence) ->
PhotoFaceTagEntity.create(imageId, faceModelId, android.graphics.Rect(0, 0, 100, 100), confidence, FloatArray(128))
}
photoFaceTagDao.insertTags(tags)
}
fun resetScanningState() { _scanningState.value = ScanningState.Idle }
fun refresh() { loadPersons() }
}
sealed class ScanningState {
object Idle : ScanningState()
data class Scanning(val personName: String, val completed: Int, val total: Int, val facesFound: Int, val speed: Double) : ScanningState()
data class Complete(val personName: String, val facesFound: Int) : ScanningState()
data class Error(val message: String) : ScanningState()
}
data class PersonWithModelInfo(val person: PersonEntity, val faceModel: FaceModelEntity?, val taggedPhotoCount: Int)

View File

@@ -1,42 +1,166 @@
// In navigation/AppDestinations.kt
package com.placeholder.sherpai2.ui.navigation package com.placeholder.sherpai2.ui.navigation
import androidx.compose.material.icons.Icons import androidx.compose.material.icons.Icons
import androidx.compose.material.icons.automirrored.filled.Label
import androidx.compose.material.icons.filled.* import androidx.compose.material.icons.filled.*
import androidx.compose.ui.graphics.vector.ImageVector import androidx.compose.ui.graphics.vector.ImageVector
/** /**
* Defines all navigation destinations (screens) for the application. * AppDestinations - Navigation metadata for drawer UI
*
* Clean, organized structure:
* - Routes for navigation
* - Icons for visual identity
* - Labels for display
* - Descriptions for clarity
* - Grouped by function
*/ */
sealed class AppDestinations(val route: String, val icon: ImageVector, val label: String) { sealed class AppDestinations(
// Core Functional Sections val route: String,
object Tour : AppDestinations( val icon: ImageVector,
route = "Tour", val label: String,
icon = Icons.Default.PhotoLibrary, val description: String = ""
label = "Tour" ) {
)
object Search : AppDestinations("search", Icons.Default.Search, "Search")
object Models : AppDestinations("models", Icons.Default.Layers, "Models")
object Inventory : AppDestinations("inv", Icons.Default.Inventory2, "Inv")
object Train : AppDestinations("train", Icons.Default.TrackChanges, "Train")
object Tags : AppDestinations("tags", Icons.Default.LocalOffer, "Tags")
// Utility/Secondary Sections // ==================
object Upload : AppDestinations("upload", Icons.Default.CloudUpload, "Upload") // PHOTO BROWSING
object Settings : AppDestinations("settings", Icons.Default.Settings, "Settings") // ==================
data object Search : AppDestinations(
route = AppRoutes.SEARCH,
icon = Icons.Default.Search,
label = "Search",
description = "Find photos by tag or person"
)
data object Explore : AppDestinations(
route = AppRoutes.EXPLORE,
icon = Icons.Default.Explore,
label = "Explore",
description = "Browse smart albums"
)
data object Collections : AppDestinations(
route = AppRoutes.COLLECTIONS,
icon = Icons.Default.Collections,
label = "Collections",
description = "Your photo collections"
)
// ImageDetail is not in drawer (internal navigation only)
// ==================
// FACE RECOGNITION
// ==================
data object Discover : AppDestinations(
route = AppRoutes.DISCOVER,
icon = Icons.Default.AutoAwesome,
label = "Discover",
description = "Find people in your photos"
)
data object Inventory : AppDestinations(
route = AppRoutes.INVENTORY,
icon = Icons.Default.Face,
label = "People",
description = "Manage recognized people"
)
data object Train : AppDestinations(
route = AppRoutes.TRAIN,
icon = Icons.Default.ModelTraining,
label = "Train Model",
description = "Create a new person model"
)
// ==================
// ORGANIZATION
// ==================
data object Tags : AppDestinations(
route = AppRoutes.TAGS,
icon = Icons.AutoMirrored.Filled.Label,
label = "Tags",
description = "Manage photo tags"
)
data object UTILITIES : AppDestinations(
route = AppRoutes.UTILITIES,
icon = Icons.Default.UploadFile,
label = "Upload",
description = "Add new photos"
)
// ==================
// SETTINGS
// ==================
data object Settings : AppDestinations(
route = AppRoutes.SETTINGS,
icon = Icons.Default.Settings,
label = "Settings",
description = "App preferences"
)
} }
// Lists used by the AppDrawerContent to render the menu sections easily /**
val mainDrawerItems = listOf( * Organized destination groups for beautiful drawer sections
AppDestinations.Tour, */
// Photo browsing section
val photoDestinations = listOf(
AppDestinations.Search, AppDestinations.Search,
AppDestinations.Models, AppDestinations.Explore,
AppDestinations.Inventory, AppDestinations.Collections
AppDestinations.Train,
AppDestinations.Tags
) )
val utilityDrawerItems = listOf( // Face recognition section
AppDestinations.Upload, val faceRecognitionDestinations = listOf(
AppDestinations.Settings AppDestinations.Discover, // ✨ NEW: Auto-cluster discovery
) AppDestinations.Inventory,
AppDestinations.Train
)
// Organization section
val organizationDestinations = listOf(
AppDestinations.Tags,
AppDestinations.UTILITIES
)
// Settings (separate, pinned to bottom)
val settingsDestination = AppDestinations.Settings
/**
* All drawer items (excludes Settings which is handled separately)
*/
val allMainDrawerDestinations = photoDestinations + faceRecognitionDestinations + organizationDestinations
/**
* Helper function to get destination by route
* Useful for highlighting current route in drawer
*/
fun getDestinationByRoute(route: String?): AppDestinations? {
return when (route) {
AppRoutes.SEARCH -> AppDestinations.Search
AppRoutes.EXPLORE -> AppDestinations.Explore
AppRoutes.COLLECTIONS -> AppDestinations.Collections
AppRoutes.DISCOVER -> AppDestinations.Discover
AppRoutes.INVENTORY -> AppDestinations.Inventory
AppRoutes.TRAIN -> AppDestinations.Train
AppRoutes.TAGS -> AppDestinations.Tags
AppRoutes.UTILITIES -> AppDestinations.UTILITIES
AppRoutes.SETTINGS -> AppDestinations.Settings
else -> null
}
}
/**
* Legacy support (for backwards compatibility)
* These match your old structure
*/
@Deprecated("Use organized groups instead", ReplaceWith("allMainDrawerDestinations"))
val mainDrawerItems = allMainDrawerDestinations
@Deprecated("Use settingsDestination instead", ReplaceWith("listOf(settingsDestination)"))
val utilityDrawerItems = listOf(settingsDestination)

View File

@@ -0,0 +1,311 @@
package com.placeholder.sherpai2.ui.navigation
import android.net.Uri
import androidx.compose.runtime.Composable
import androidx.compose.runtime.LaunchedEffect
import androidx.compose.runtime.collectAsState
import androidx.compose.runtime.getValue
import androidx.compose.ui.Modifier
import androidx.hilt.navigation.compose.hiltViewModel
import androidx.lifecycle.compose.collectAsStateWithLifecycle
import androidx.navigation.NavHostController
import androidx.navigation.NavType
import androidx.navigation.compose.NavHost
import androidx.navigation.compose.composable
import androidx.navigation.navArgument
import com.placeholder.sherpai2.ui.devscreens.DummyScreen
import com.placeholder.sherpai2.ui.album.AlbumViewScreen
import com.placeholder.sherpai2.ui.album.AlbumViewModel
import com.placeholder.sherpai2.ui.collections.CollectionsScreen
import com.placeholder.sherpai2.ui.collections.CollectionsViewModel
import com.placeholder.sherpai2.ui.discover.DiscoverPeopleScreen
import com.placeholder.sherpai2.ui.explore.ExploreScreen
import com.placeholder.sherpai2.ui.imagedetail.ImageDetailScreen
import com.placeholder.sherpai2.ui.modelinventory.PersonInventoryScreen
import com.placeholder.sherpai2.ui.search.SearchScreen
import com.placeholder.sherpai2.ui.search.SearchViewModel
import com.placeholder.sherpai2.ui.tags.TagManagementScreen
import com.placeholder.sherpai2.ui.trainingprep.ScanResultsScreen
import com.placeholder.sherpai2.ui.trainingprep.ScanningState
import com.placeholder.sherpai2.ui.trainingprep.TrainViewModel
import com.placeholder.sherpai2.ui.trainingprep.TrainingScreen
import com.placeholder.sherpai2.ui.trainingprep.TrainingPhotoSelectorScreen
import com.placeholder.sherpai2.ui.utilities.PhotoUtilitiesScreen
import java.net.URLDecoder
import java.net.URLEncoder
import com.placeholder.sherpai2.ui.navigation.AppRoutes
/**
* AppNavHost - UPDATED with Discover People screen
*
* NEW: Replaces placeholder "Models" screen with auto-clustering face discovery
*/
@Composable
fun AppNavHost(
navController: NavHostController,
modifier: Modifier = Modifier
) {
NavHost(
navController = navController,
startDestination = AppRoutes.SEARCH,
modifier = modifier
) {
// ==========================================
// PHOTO BROWSING
// ==========================================
/**
* SEARCH SCREEN
*/
composable(AppRoutes.SEARCH) {
val searchViewModel: SearchViewModel = hiltViewModel()
val collectionsViewModel: CollectionsViewModel = hiltViewModel()
SearchScreen(
searchViewModel = searchViewModel,
onImageClick = { imageUri ->
ImageListHolder.clear()
val encodedUri = URLEncoder.encode(imageUri, "UTF-8")
navController.navigate("${AppRoutes.IMAGE_DETAIL}/$encodedUri")
},
onAlbumClick = { tagValue ->
navController.navigate("album/tag/$tagValue")
},
onSaveToCollection = { includedPeople, excludedPeople, includedTags, excludedTags, dateRange, photoCount ->
collectionsViewModel.startSmartCollectionFromSearch(
includedPeople = includedPeople,
excludedPeople = excludedPeople,
includedTags = includedTags,
excludedTags = excludedTags,
dateRange = dateRange,
photoCount = photoCount
)
}
)
}
/**
* EXPLORE SCREEN
*/
composable(AppRoutes.EXPLORE) {
ExploreScreen(
onAlbumClick = { albumType, albumId ->
navController.navigate("album/$albumType/$albumId")
}
)
}
/**
* COLLECTIONS SCREEN
*/
composable(AppRoutes.COLLECTIONS) {
val collectionsViewModel: CollectionsViewModel = hiltViewModel()
CollectionsScreen(
viewModel = collectionsViewModel,
onCollectionClick = { collectionId ->
navController.navigate("album/collection/$collectionId")
},
onCreateClick = {
navController.navigate(AppRoutes.SEARCH)
}
)
}
/**
* IMAGE DETAIL SCREEN
*/
composable(
route = "${AppRoutes.IMAGE_DETAIL}/{imageUri}",
arguments = listOf(
navArgument("imageUri") {
type = NavType.StringType
}
)
) { backStackEntry ->
val imageUri = backStackEntry.arguments?.getString("imageUri")
?.let { URLDecoder.decode(it, "UTF-8") }
?: error("imageUri missing from navigation")
val allImageUris = ImageListHolder.getImageList()
ImageDetailScreen(
imageUri = imageUri,
onBack = {
ImageListHolder.clear()
navController.popBackStack()
},
navController = navController,
allImageUris = allImageUris
)
}
/**
* ALBUM VIEW SCREEN
*/
composable(
route = "album/{albumType}/{albumId}",
arguments = listOf(
navArgument("albumType") {
type = NavType.StringType
},
navArgument("albumId") {
type = NavType.StringType
}
)
) {
val albumViewModel: AlbumViewModel = hiltViewModel()
val uiState by albumViewModel.uiState.collectAsStateWithLifecycle()
AlbumViewScreen(
onBack = {
navController.popBackStack()
},
onImageClick = { imageUri ->
val allImageUris = if (uiState is com.placeholder.sherpai2.ui.album.AlbumUiState.Success) {
(uiState as com.placeholder.sherpai2.ui.album.AlbumUiState.Success)
.photos
.map { it.image.imageUri }
} else {
emptyList()
}
ImageListHolder.setImageList(allImageUris)
val encodedUri = URLEncoder.encode(imageUri, "UTF-8")
navController.navigate("${AppRoutes.IMAGE_DETAIL}/$encodedUri")
}
)
}
// ==========================================
// FACE RECOGNITION SYSTEM
// ==========================================
/**
* DISCOVER PEOPLE SCREEN - ✨ NEW!
*
* Auto-clustering face discovery with spoon-feed naming flow:
* 1. Auto-clusters all faces in library (2-5 min)
* 2. Shows beautiful grid of discovered people
* 3. User taps to name each person
* 4. Captures: name, DOB, sibling relationships
* 5. Triggers deep background scan with age tagging
*
* Replaces: Old "Models" placeholder screen
*/
composable(AppRoutes.DISCOVER) {
DiscoverPeopleScreen()
}
/**
* PERSON INVENTORY SCREEN
*/
composable(AppRoutes.INVENTORY) {
PersonInventoryScreen(
onNavigateToPersonDetail = { personId ->
navController.navigate(AppRoutes.SEARCH)
}
)
}
/**
* TRAINING FLOW - Manual training (still available)
*/
composable(AppRoutes.TRAIN) { entry ->
val trainViewModel: TrainViewModel = hiltViewModel()
val uiState by trainViewModel.uiState.collectAsState()
val selectedUris = entry.savedStateHandle.get<List<Uri>>("selected_image_uris")
LaunchedEffect(selectedUris) {
if (selectedUris != null && uiState is ScanningState.Idle) {
trainViewModel.scanAndTagFaces(selectedUris)
entry.savedStateHandle.remove<List<Uri>>("selected_image_uris")
}
}
when (uiState) {
is ScanningState.Idle -> {
TrainingScreen(
onSelectImages = {
// Navigate to custom photo selector (shows only faces!)
navController.navigate(AppRoutes.TRAINING_PHOTO_SELECTOR)
}
)
}
else -> {
ScanResultsScreen(
state = uiState,
onFinish = {
navController.navigate(AppRoutes.INVENTORY) {
popUpTo(AppRoutes.TRAIN) { inclusive = true }
}
}
)
}
}
}
/**
* TRAINING PHOTO SELECTOR - Custom gallery with face filtering
*/
composable(AppRoutes.TRAINING_PHOTO_SELECTOR) {
TrainingPhotoSelectorScreen(
onBack = {
navController.popBackStack()
},
onPhotosSelected = { uris ->
// Pass selected URIs back to training flow
navController.previousBackStackEntry
?.savedStateHandle
?.set("selected_image_uris", uris)
navController.popBackStack()
}
)
}
/**
* MODELS SCREEN - DEPRECATED, kept for backwards compat
*/
composable(AppRoutes.MODELS) {
DummyScreen(
title = "AI Models",
subtitle = "Use 'Discover' instead"
)
}
// ==========================================
// ORGANIZATION
// ==========================================
/**
* TAGS SCREEN
*/
composable(AppRoutes.TAGS) {
TagManagementScreen()
}
/**
* UTILITIES SCREEN
*/
composable(AppRoutes.UTILITIES) {
PhotoUtilitiesScreen()
}
// ==========================================
// SETTINGS
// ==========================================
/**
* SETTINGS SCREEN
*/
composable(AppRoutes.SETTINGS) {
DummyScreen(
title = "Settings",
subtitle = "App preferences and configuration"
)
}
}
}

View File

@@ -0,0 +1,45 @@
package com.placeholder.sherpai2.ui.navigation
/**
* Centralized list of navigation routes used by NavHost.
*
* This intentionally mirrors AppDestinations.route
* but exists as a pure navigation concern.
*
* Why:
* - Drawer UI ≠ Navigation system
* - Keeps NavHost decoupled from icons / labels
*/
object AppRoutes {
// Photo browsing
const val SEARCH = "search"
const val EXPLORE = "explore"
const val IMAGE_DETAIL = "IMAGE_DETAIL"
// Face recognition
const val DISCOVER = "discover" // ✨ NEW: Auto-cluster face discovery
const val INVENTORY = "inv"
const val TRAIN = "train"
const val MODELS = "models" // DEPRECATED - kept for reference only
// Organization
const val TAGS = "tags"
const val UTILITIES = "utilities"
// Settings
const val SETTINGS = "settings"
// Internal training flow screens
const val IMAGE_SELECTOR = "Image Selection" // DEPRECATED - kept for reference only
const val TRAINING_PHOTO_SELECTOR = "training_photo_selector" // Face-filtered gallery
const val CROP_SCREEN = "CROP_SCREEN"
const val TRAINING_SCREEN = "TRAINING_SCREEN"
const val ScanResultsScreen = "First Scan Results"
// Album view
const val ALBUM_VIEW = "album/{albumType}/{albumId}"
fun albumRoute(albumType: String, albumId: String) = "album/$albumType/$albumId"
// Collections
const val COLLECTIONS = "collections"
}

View File

@@ -0,0 +1,21 @@
package com.placeholder.sherpai2.ui.navigation
/**
* Simple holder for passing image lists between screens
* Used for prev/next navigation in ImageDetailScreen
*/
object ImageListHolder {
private var imageUris: List<String> = emptyList()
fun setImageList(uris: List<String>) {
imageUris = uris
}
fun getImageList(): List<String> {
return imageUris
}
fun clear() {
imageUris = emptyList()
}
}

View File

@@ -1,63 +0,0 @@
// In presentation/MainScreen.kt
package com.placeholder.sherpai2.presentation
import GalleryViewModel
import androidx.compose.foundation.layout.padding
import androidx.compose.material.icons.Icons
import androidx.compose.material.icons.filled.Menu
import androidx.compose.material3.*
import androidx.compose.runtime.*
import androidx.compose.ui.Modifier
import androidx.lifecycle.viewmodel.compose.viewModel
import com.placeholder.sherpai2.ui.navigation.AppDestinations
import com.placeholder.sherpai2.ui.presentation.AppDrawerContent
import com.placeholder.sherpai2.ui.presentation.MainContentArea
import kotlinx.coroutines.launch
@OptIn(ExperimentalMaterial3Api::class)
@Composable
fun MainScreen() {
val drawerState = rememberDrawerState(initialValue = DrawerValue.Closed)
val scope = rememberCoroutineScope()
// State to track which screen is currently visible
// var currentScreen by remember { mutableStateOf(AppDestinations.Search) }
var currentScreen: AppDestinations by remember { mutableStateOf(AppDestinations.Search) }
val galleryViewModel: GalleryViewModel = viewModel()
// ModalNavigationDrawer provides the left sidebar UI/UX
ModalNavigationDrawer(
drawerState = drawerState,
drawerContent = {
// The content of the drawer (AppDrawerContent)
AppDrawerContent(
currentScreen = currentScreen,
onDestinationClicked = { destination ->
currentScreen = destination
scope.launch { drawerState.close() } // Close drawer after selection
}
)
},
) {
// The main content area
Scaffold(
topBar = {
TopAppBar(
title = { Text(currentScreen.label) },
// Button to open the drawer
navigationIcon = {
IconButton(onClick = { scope.launch { drawerState.open() } }) {
Icon(Icons.Filled.Menu, contentDescription = "Open Drawer")
}
}
)
}
) { paddingValues ->
// Displays the content for the currently selected screen
MainContentArea(
currentScreen = currentScreen,
galleryViewModel = galleryViewModel,
modifier = Modifier.padding(paddingValues)
)
}
}
}

View File

@@ -1,59 +1,243 @@
// In presentation/AppDrawerContent.kt
package com.placeholder.sherpai2.ui.presentation package com.placeholder.sherpai2.ui.presentation
import androidx.compose.foundation.background
import androidx.compose.foundation.layout.* import androidx.compose.foundation.layout.*
import androidx.compose.foundation.rememberScrollState
import androidx.compose.foundation.shape.RoundedCornerShape
import androidx.compose.foundation.verticalScroll
import androidx.compose.material3.* import androidx.compose.material3.*
import androidx.compose.runtime.Composable import androidx.compose.runtime.Composable
import androidx.compose.ui.Alignment
import androidx.compose.ui.Modifier import androidx.compose.ui.Modifier
import androidx.compose.ui.draw.clip
import androidx.compose.ui.graphics.Brush
import androidx.compose.ui.graphics.Color
import androidx.compose.ui.text.font.FontWeight
import androidx.compose.ui.unit.dp import androidx.compose.ui.unit.dp
import com.placeholder.sherpai2.ui.navigation.AppDestinations import androidx.compose.material.icons.Icons
import com.placeholder.sherpai2.ui.navigation.mainDrawerItems import androidx.compose.material.icons.automirrored.filled.Label
import com.placeholder.sherpai2.ui.navigation.utilityDrawerItems import androidx.compose.material.icons.filled.*
import com.placeholder.sherpai2.ui.navigation.AppRoutes
/**
* SLIMMED DOWN AppDrawer - 280dp width, inline logo, cleaner sections
* UPDATED: Discover People feature with sparkle icon ✨
*/
@OptIn(ExperimentalMaterial3Api::class) @OptIn(ExperimentalMaterial3Api::class)
@Composable @Composable
fun AppDrawerContent( fun AppDrawerContent(
currentScreen: AppDestinations, currentRoute: String?,
onDestinationClicked: (AppDestinations) -> Unit onDestinationClicked: (String) -> Unit
) { ) {
// Defines the width and content of the sliding drawer panel ModalDrawerSheet(
ModalDrawerSheet(modifier = Modifier.width(280.dp)) { modifier = Modifier.width(280.dp), // SLIMMER (was 300dp)
drawerContainerColor = MaterialTheme.colorScheme.surface
) {
// SCROLLABLE Column - works on small phones!
Column(
modifier = Modifier
.fillMaxSize()
.verticalScroll(rememberScrollState())
) {
// Header/Logo Area // ===== COMPACT HEADER - Icon + Text Inline =====
Text( Box(
"SherpAI Control Panel", modifier = Modifier
style = MaterialTheme.typography.headlineSmall, .fillMaxWidth()
modifier = Modifier.padding(16.dp) .background(
) Brush.verticalGradient(
Divider(Modifier.fillMaxWidth()) colors = listOf(
MaterialTheme.colorScheme.primaryContainer,
MaterialTheme.colorScheme.surface
)
)
)
.padding(20.dp) // Reduced padding
) {
Row(
horizontalArrangement = Arrangement.spacedBy(12.dp),
verticalAlignment = Alignment.CenterVertically
) {
// App icon - smaller
Surface(
modifier = Modifier.size(48.dp), // Smaller (was 56dp)
shape = RoundedCornerShape(14.dp),
color = MaterialTheme.colorScheme.primary,
shadowElevation = 4.dp
) {
Box(contentAlignment = Alignment.Center) {
Icon(
Icons.Default.Terrain, // Mountain theme!
contentDescription = null,
modifier = Modifier.size(28.dp),
tint = MaterialTheme.colorScheme.onPrimary
)
}
}
// 1. Main Navigation Items // Text next to icon
Column(modifier = Modifier.padding(vertical = 8.dp)) { Column(verticalArrangement = Arrangement.spacedBy(2.dp)) {
mainDrawerItems.forEach { destination -> Text(
NavigationDrawerItem( "SherpAI",
label = { Text(destination.label) }, style = MaterialTheme.typography.titleLarge, // Smaller (was headlineMedium)
icon = { Icon(destination.icon, contentDescription = destination.label) }, fontWeight = FontWeight.Bold,
selected = destination == currentScreen, color = MaterialTheme.colorScheme.onSurface
onClick = { onDestinationClicked(destination) }, )
modifier = Modifier.padding(NavigationDrawerItemDefaults.ItemPadding)
) Text(
"Face Recognition",
style = MaterialTheme.typography.bodySmall, // Smaller
color = MaterialTheme.colorScheme.onSurfaceVariant
)
}
}
} }
}
// Separator Spacer(modifier = Modifier.height(4.dp)) // Reduced spacing
Divider(Modifier.fillMaxWidth().padding(vertical = 8.dp))
// 2. Utility Items // ===== NAVIGATION SECTIONS =====
Column(modifier = Modifier.padding(vertical = 8.dp)) { Column(
utilityDrawerItems.forEach { destination -> modifier = Modifier
NavigationDrawerItem( .fillMaxWidth()
label = { Text(destination.label) }, .padding(horizontal = 8.dp), // Reduced padding
icon = { Icon(destination.icon, contentDescription = destination.label) }, verticalArrangement = Arrangement.spacedBy(2.dp) // Tighter spacing
selected = destination == currentScreen, ) {
onClick = { onDestinationClicked(destination) },
modifier = Modifier.padding(NavigationDrawerItemDefaults.ItemPadding) // Photos Section
DrawerSection(title = "Photos")
val photoItems = listOf(
DrawerItem(AppRoutes.SEARCH, "Search", Icons.Default.Search),
DrawerItem(AppRoutes.EXPLORE, "Explore", Icons.Default.Explore),
DrawerItem(AppRoutes.COLLECTIONS, "Collections", Icons.Default.Collections)
) )
photoItems.forEach { item ->
DrawerNavigationItem(
item = item,
selected = item.route == currentRoute,
onClick = { onDestinationClicked(item.route) }
)
}
Spacer(modifier = Modifier.height(4.dp))
// Face Recognition Section
DrawerSection(title = "Face Recognition")
val faceItems = listOf(
DrawerItem(AppRoutes.DISCOVER, "Discover", Icons.Default.AutoAwesome), // ✨ UPDATED!
DrawerItem(AppRoutes.INVENTORY, "People", Icons.Default.Face),
DrawerItem(AppRoutes.TRAIN, "Train Model", Icons.Default.ModelTraining)
)
faceItems.forEach { item ->
DrawerNavigationItem(
item = item,
selected = item.route == currentRoute,
onClick = { onDestinationClicked(item.route) }
)
}
Spacer(modifier = Modifier.height(4.dp))
// Organization Section
DrawerSection(title = "Organization")
val orgItems = listOf(
DrawerItem(AppRoutes.TAGS, "Tags", Icons.AutoMirrored.Filled.Label),
DrawerItem(AppRoutes.UTILITIES, "Utilities", Icons.Default.Build)
)
orgItems.forEach { item ->
DrawerNavigationItem(
item = item,
selected = item.route == currentRoute,
onClick = { onDestinationClicked(item.route) }
)
}
Spacer(modifier = Modifier.height(8.dp))
// Settings at bottom
HorizontalDivider(
modifier = Modifier.padding(vertical = 6.dp),
color = MaterialTheme.colorScheme.outlineVariant
)
DrawerNavigationItem(
item = DrawerItem(
AppRoutes.SETTINGS,
"Settings",
Icons.Default.Settings
),
selected = AppRoutes.SETTINGS == currentRoute,
onClick = { onDestinationClicked(AppRoutes.SETTINGS) }
)
Spacer(modifier = Modifier.height(16.dp)) // Bottom padding for scroll
} }
} }
} }
} }
/**
* Section header - more compact
*/
@Composable
private fun DrawerSection(title: String) {
Text(
text = title,
style = MaterialTheme.typography.labelSmall, // Smaller
fontWeight = FontWeight.Bold,
color = MaterialTheme.colorScheme.primary,
modifier = Modifier.padding(horizontal = 16.dp, vertical = 6.dp) // Reduced padding
)
}
/**
* Navigation item - cleaner, no subtitle
*/
@Composable
private fun DrawerNavigationItem(
item: DrawerItem,
selected: Boolean,
onClick: () -> Unit
) {
NavigationDrawerItem(
label = {
Text(
text = item.label,
style = MaterialTheme.typography.bodyMedium, // Slightly smaller
fontWeight = if (selected) FontWeight.SemiBold else FontWeight.Normal
)
},
icon = {
Icon(
item.icon,
contentDescription = item.label,
modifier = Modifier.size(22.dp) // Slightly smaller
)
},
selected = selected,
onClick = onClick,
modifier = Modifier
.padding(NavigationDrawerItemDefaults.ItemPadding)
.clip(RoundedCornerShape(10.dp)), // Slightly smaller radius
colors = NavigationDrawerItemDefaults.colors(
selectedContainerColor = MaterialTheme.colorScheme.primaryContainer,
selectedIconColor = MaterialTheme.colorScheme.primary,
selectedTextColor = MaterialTheme.colorScheme.onPrimaryContainer,
unselectedContainerColor = Color.Transparent
)
)
}
/**
* Simplified drawer item (no subtitle)
*/
private data class DrawerItem(
val route: String,
val label: String,
val icon: androidx.compose.ui.graphics.vector.ImageVector
)

View File

@@ -1,56 +0,0 @@
// In presentation/MainContentArea.kt
package com.placeholder.sherpai2.ui.presentation
import GalleryScreen
import GalleryViewModel
import androidx.compose.ui.graphics.Color
import androidx.compose.foundation.background
import androidx.compose.foundation.layout.*
import androidx.compose.foundation.shape.RoundedCornerShape
import androidx.compose.material3.MaterialTheme
import androidx.compose.material3.Text
import androidx.compose.runtime.Composable
import androidx.compose.runtime.collectAsState
import androidx.compose.runtime.getValue
import androidx.compose.ui.Alignment
import androidx.compose.ui.Modifier
import androidx.compose.ui.text.font.FontWeight
import androidx.compose.ui.tooling.preview.Preview
import androidx.compose.ui.unit.dp
import com.placeholder.sherpai2.ui.navigation.AppDestinations
import com.placeholder.sherpai2.ui.theme.SherpAI2Theme
import androidx.lifecycle.viewmodel.compose.viewModel
@Composable
fun MainContentArea(currentScreen: AppDestinations, modifier: Modifier = Modifier, galleryViewModel: GalleryViewModel = viewModel() ) {
val uiState by galleryViewModel.uiState.collectAsState()
Box(
modifier = modifier
.fillMaxSize()
.background(Color.Red),
contentAlignment = Alignment.Center
) {
// Swaps the UI content based on the selected screen from the drawer
when (currentScreen) {
AppDestinations.Tour -> GalleryScreen(state = uiState, modifier = Modifier)
AppDestinations.Search -> SimplePlaceholder("Find Any Photos.")
AppDestinations.Models -> SimplePlaceholder("Models Screen: Manage your LoRA/embeddings.")
AppDestinations.Inventory -> SimplePlaceholder("Inventory Screen: View all collected data.")
AppDestinations.Train -> SimplePlaceholder("Train Screen: Start the LoRA adaptation process.")
AppDestinations.Tags -> SimplePlaceholder("Tags Screen: Create and edit custom tags.")
AppDestinations.Upload -> SimplePlaceholder("Upload Screen: Import new photos/data.")
AppDestinations.Settings -> SimplePlaceholder("Settings Screen: Configure app behavior.")
}
}
}
@Composable
private fun SimplePlaceholder(text: String) {
Text(
text = text,
style = MaterialTheme.typography.titleLarge,
modifier = Modifier.padding(16.dp).background(color = Color.Magenta)
)
}

View File

@@ -0,0 +1,154 @@
package com.placeholder.sherpai2.ui.presentation
import androidx.compose.foundation.layout.Column
import androidx.compose.foundation.layout.padding
import androidx.compose.material.icons.Icons
import androidx.compose.material.icons.filled.*
import androidx.compose.material3.*
import androidx.compose.runtime.*
import androidx.compose.ui.Modifier
import androidx.compose.ui.text.font.FontWeight
import androidx.navigation.compose.currentBackStackEntryAsState
import androidx.navigation.compose.rememberNavController
import com.placeholder.sherpai2.ui.navigation.AppNavHost
import com.placeholder.sherpai2.ui.navigation.AppRoutes
import kotlinx.coroutines.launch
/**
* Clean main screen - NO duplicate FABs, Collections support, Discover People
*/
@OptIn(ExperimentalMaterial3Api::class)
@Composable
fun MainScreen() {
val drawerState = rememberDrawerState(initialValue = DrawerValue.Closed)
val scope = rememberCoroutineScope()
val navController = rememberNavController()
val navBackStackEntry by navController.currentBackStackEntryAsState()
val currentRoute = navBackStackEntry?.destination?.route ?: AppRoutes.SEARCH
ModalNavigationDrawer(
drawerState = drawerState,
drawerContent = {
AppDrawerContent(
currentRoute = currentRoute,
onDestinationClicked = { route ->
scope.launch {
drawerState.close()
if (route != currentRoute) {
navController.navigate(route) {
launchSingleTop = true
}
}
}
}
)
},
) {
Scaffold(
topBar = {
TopAppBar(
title = {
Column {
Text(
text = getScreenTitle(currentRoute),
style = MaterialTheme.typography.titleLarge,
fontWeight = FontWeight.Bold
)
getScreenSubtitle(currentRoute)?.let { subtitle ->
Text(
text = subtitle,
style = MaterialTheme.typography.bodySmall,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
}
}
},
navigationIcon = {
IconButton(
onClick = { scope.launch { drawerState.open() } }
) {
Icon(
Icons.Default.Menu,
contentDescription = "Open Menu",
tint = MaterialTheme.colorScheme.primary
)
}
},
actions = {
// Dynamic actions based on current screen
when (currentRoute) {
AppRoutes.SEARCH -> {
IconButton(onClick = { /* TODO: Open filter dialog */ }) {
Icon(
Icons.Default.FilterList,
contentDescription = "Filter",
tint = MaterialTheme.colorScheme.primary
)
}
}
AppRoutes.INVENTORY -> {
IconButton(onClick = {
navController.navigate(AppRoutes.TRAIN)
}) {
Icon(
Icons.Default.PersonAdd,
contentDescription = "Add Person",
tint = MaterialTheme.colorScheme.primary
)
}
}
}
},
colors = TopAppBarDefaults.topAppBarColors(
containerColor = MaterialTheme.colorScheme.surface,
titleContentColor = MaterialTheme.colorScheme.onSurface,
navigationIconContentColor = MaterialTheme.colorScheme.primary,
actionIconContentColor = MaterialTheme.colorScheme.primary
)
)
}
) { paddingValues ->
AppNavHost(
navController = navController,
modifier = Modifier.padding(paddingValues)
)
}
}
}
/**
* Get human-readable screen title
*/
private fun getScreenTitle(route: String): String {
return when (route) {
AppRoutes.SEARCH -> "Search"
AppRoutes.EXPLORE -> "Explore"
AppRoutes.COLLECTIONS -> "Collections"
AppRoutes.DISCOVER -> "Discover People" // ✨ NEW!
AppRoutes.INVENTORY -> "People"
AppRoutes.TRAIN -> "Train New Person"
AppRoutes.MODELS -> "AI Models" // Deprecated, but keep for backwards compat
AppRoutes.TAGS -> "Tag Management"
AppRoutes.UTILITIES -> "Photo Util."
AppRoutes.SETTINGS -> "Settings"
else -> "SherpAI"
}
}
/**
* Get subtitle for screens that need context
*/
private fun getScreenSubtitle(route: String): String? {
return when (route) {
AppRoutes.SEARCH -> "Find photos by tags, people, or date"
AppRoutes.EXPLORE -> "Browse your collection"
AppRoutes.COLLECTIONS -> "Your photo collections"
AppRoutes.DISCOVER -> "Auto-find faces in your library" // ✨ NEW!
AppRoutes.INVENTORY -> "Trained face models"
AppRoutes.TRAIN -> "Add a new person to recognize"
AppRoutes.TAGS -> "Organize your photo collection"
AppRoutes.UTILITIES -> "Tools for managing collection"
else -> null
}
}

View File

@@ -1,34 +0,0 @@
package com.placeholder.sherpai2.ui.presentation
import androidx.compose.foundation.Image
import androidx.compose.foundation.layout.*
import androidx.compose.foundation.lazy.LazyColumn
import androidx.compose.foundation.lazy.items
import androidx.compose.runtime.Composable
import androidx.compose.ui.Modifier
import androidx.compose.ui.unit.dp
import coil.compose.rememberAsyncImagePainter
import com.placeholder.sherpai2.data.photos.Photo
@Composable
fun PhotoListScreen(
photos: List<Photo>
) {
LazyColumn(
modifier = Modifier.fillMaxSize(),
contentPadding = PaddingValues(8.dp)
) {
items(photos, key = { it.id }) { photo ->
Image(
painter = rememberAsyncImagePainter(photo.uri),
contentDescription = photo.title,
modifier = Modifier
.fillMaxWidth()
.height(200.dp)
.padding(bottom = 8.dp)
)
}
}
}

View File

@@ -0,0 +1,728 @@
package com.placeholder.sherpai2.ui.search
import androidx.compose.foundation.clickable
import androidx.compose.foundation.layout.*
import androidx.compose.foundation.lazy.LazyRow
import androidx.compose.foundation.lazy.grid.*
import androidx.compose.foundation.lazy.items
import androidx.compose.foundation.rememberScrollState
import androidx.compose.foundation.shape.RoundedCornerShape
import androidx.compose.foundation.verticalScroll
import androidx.compose.material.icons.Icons
import androidx.compose.material.icons.filled.*
import androidx.compose.material3.*
import androidx.compose.runtime.*
import androidx.compose.ui.Alignment
import androidx.compose.ui.Modifier
import androidx.compose.ui.graphics.Color
import androidx.compose.ui.text.font.FontWeight
import androidx.compose.ui.unit.dp
import androidx.lifecycle.compose.collectAsStateWithLifecycle
import coil.compose.AsyncImage
import com.placeholder.sherpai2.data.local.entity.PersonEntity
/**
* ENHANCED SearchScreen
*
* NEW FEATURES:
* ✅ Face filtering (Has Faces / No Faces)
* ✅ X button on each filter chip for easy removal
* ✅ Tap to swap include/exclude (kept)
* ✅ Better visual hierarchy
*/
@OptIn(ExperimentalMaterial3Api::class)
@Composable
fun SearchScreen(
modifier: Modifier = Modifier,
searchViewModel: SearchViewModel,
onImageClick: (String) -> Unit,
onAlbumClick: ((String) -> Unit)? = null,
onSaveToCollection: ((
includedPeople: Set<String>,
excludedPeople: Set<String>,
includedTags: Set<String>,
excludedTags: Set<String>,
dateRange: DateRange,
photoCount: Int
) -> Unit)? = null
) {
val searchQuery by searchViewModel.searchQuery.collectAsStateWithLifecycle()
val includedPeople by searchViewModel.includedPeople.collectAsStateWithLifecycle()
val excludedPeople by searchViewModel.excludedPeople.collectAsStateWithLifecycle()
val includedTags by searchViewModel.includedTags.collectAsStateWithLifecycle()
val excludedTags by searchViewModel.excludedTags.collectAsStateWithLifecycle()
val dateRange by searchViewModel.dateRange.collectAsStateWithLifecycle()
val faceFilter by searchViewModel.faceFilter.collectAsStateWithLifecycle()
val availablePeople by searchViewModel.availablePeople.collectAsStateWithLifecycle()
val availableTags by searchViewModel.availableTags.collectAsStateWithLifecycle()
val images by searchViewModel
.searchImages()
.collectAsStateWithLifecycle(initialValue = emptyList())
var showPeoplePicker by remember { mutableStateOf(false) }
var showTagPicker by remember { mutableStateOf(false) }
var showFaceFilterMenu by remember { mutableStateOf(false) }
Column(modifier = modifier.fillMaxSize()) {
// Search bar + quick add buttons
Row(
modifier = Modifier
.fillMaxWidth()
.padding(horizontal = 16.dp, vertical = 12.dp),
horizontalArrangement = Arrangement.spacedBy(8.dp),
verticalAlignment = Alignment.CenterVertically
) {
OutlinedTextField(
value = searchQuery,
onValueChange = { searchViewModel.setSearchQuery(it) },
placeholder = { Text("Search tags...") },
leadingIcon = { Icon(Icons.Default.Search, null) },
trailingIcon = {
if (searchQuery.isNotEmpty()) {
IconButton(onClick = { searchViewModel.setSearchQuery("") }) {
Icon(Icons.Default.Close, "Clear")
}
}
},
modifier = Modifier.weight(1f),
singleLine = true,
shape = RoundedCornerShape(12.dp)
)
// Add person button
IconButton(
onClick = { showPeoplePicker = true },
colors = IconButtonDefaults.iconButtonColors(
containerColor = MaterialTheme.colorScheme.primaryContainer
)
) {
Icon(Icons.Default.PersonAdd, "Add person filter")
}
// Add tag button
IconButton(
onClick = { showTagPicker = true },
colors = IconButtonDefaults.iconButtonColors(
containerColor = MaterialTheme.colorScheme.secondaryContainer
)
) {
Icon(Icons.Default.LabelImportant, "Add tag filter")
}
// Face filter button (NEW!)
IconButton(
onClick = { showFaceFilterMenu = true },
colors = IconButtonDefaults.iconButtonColors(
containerColor = if (faceFilter != FaceFilter.ALL) {
MaterialTheme.colorScheme.tertiaryContainer
} else {
MaterialTheme.colorScheme.surfaceVariant
}
)
) {
Icon(
when (faceFilter) {
FaceFilter.HAS_FACES -> Icons.Default.Face
FaceFilter.NO_FACES -> Icons.Default.HideImage
else -> Icons.Default.FilterAlt
},
"Face filter"
)
}
}
// Active filters display (chips)
if (searchViewModel.hasActiveFilters()) {
Card(
modifier = Modifier
.fillMaxWidth()
.padding(horizontal = 16.dp, vertical = 4.dp),
colors = CardDefaults.cardColors(
containerColor = MaterialTheme.colorScheme.surfaceVariant.copy(alpha = 0.5f)
)
) {
Column(
modifier = Modifier.padding(12.dp),
verticalArrangement = Arrangement.spacedBy(8.dp)
) {
Row(
modifier = Modifier.fillMaxWidth(),
horizontalArrangement = Arrangement.SpaceBetween,
verticalAlignment = Alignment.CenterVertically
) {
Text(
"Active Filters",
style = MaterialTheme.typography.labelLarge,
fontWeight = FontWeight.Bold
)
Row(horizontalArrangement = Arrangement.spacedBy(8.dp)) {
// Save to Collection button
if (onSaveToCollection != null && images.isNotEmpty()) {
FilledTonalButton(
onClick = {
onSaveToCollection(
includedPeople,
excludedPeople,
includedTags,
excludedTags,
dateRange,
images.size
)
},
contentPadding = PaddingValues(horizontal = 12.dp, vertical = 4.dp)
) {
Icon(
Icons.Default.Collections,
contentDescription = null,
modifier = Modifier.size(16.dp)
)
Spacer(Modifier.width(4.dp))
Text("Save", style = MaterialTheme.typography.labelMedium)
}
}
TextButton(
onClick = { searchViewModel.clearAllFilters() },
contentPadding = PaddingValues(horizontal = 8.dp, vertical = 4.dp)
) {
Text("Clear All", style = MaterialTheme.typography.labelMedium)
}
}
}
// Face Filter Chip (NEW!)
if (faceFilter != FaceFilter.ALL) {
FilterChipWithX(
label = faceFilter.displayName,
color = MaterialTheme.colorScheme.tertiaryContainer,
onTap = { showFaceFilterMenu = true },
onRemove = { searchViewModel.setFaceFilter(FaceFilter.ALL) },
leadingIcon = {
Icon(
when (faceFilter) {
FaceFilter.HAS_FACES -> Icons.Default.Face
FaceFilter.NO_FACES -> Icons.Default.HideImage
else -> Icons.Default.FilterAlt
},
contentDescription = null,
modifier = Modifier.size(16.dp)
)
}
)
}
// Included People (GREEN)
if (includedPeople.isNotEmpty()) {
LazyRow(
horizontalArrangement = Arrangement.spacedBy(6.dp),
contentPadding = PaddingValues(vertical = 4.dp)
) {
items(includedPeople.toList()) { personId ->
val person = availablePeople.find { it.id == personId }
if (person != null) {
FilterChipWithX(
label = person.name,
color = Color(0xFF4CAF50).copy(alpha = 0.3f),
onTap = { searchViewModel.excludePerson(personId) },
onRemove = { searchViewModel.removePersonFilter(personId) },
leadingIcon = {
Icon(
Icons.Default.Person,
contentDescription = null,
modifier = Modifier.size(16.dp),
tint = Color(0xFF2E7D32)
)
}
)
}
}
}
}
// Excluded People (RED)
if (excludedPeople.isNotEmpty()) {
LazyRow(
horizontalArrangement = Arrangement.spacedBy(6.dp),
contentPadding = PaddingValues(vertical = 4.dp)
) {
items(excludedPeople.toList()) { personId ->
val person = availablePeople.find { it.id == personId }
if (person != null) {
FilterChipWithX(
label = person.name,
color = Color(0xFFF44336).copy(alpha = 0.3f),
onTap = { searchViewModel.includePerson(personId) },
onRemove = { searchViewModel.removePersonFilter(personId) },
leadingIcon = {
Icon(
Icons.Default.PersonOff,
contentDescription = null,
modifier = Modifier.size(16.dp),
tint = Color(0xFFC62828)
)
}
)
}
}
}
}
// Included Tags (GREEN)
if (includedTags.isNotEmpty()) {
LazyRow(
horizontalArrangement = Arrangement.spacedBy(6.dp),
contentPadding = PaddingValues(vertical = 4.dp)
) {
items(includedTags.toList()) { tag ->
FilterChipWithX(
label = tag,
color = Color(0xFF4CAF50).copy(alpha = 0.3f),
onTap = { searchViewModel.excludeTag(tag) },
onRemove = { searchViewModel.removeTagFilter(tag) },
leadingIcon = {
Icon(
Icons.Default.Label,
contentDescription = null,
modifier = Modifier.size(16.dp),
tint = Color(0xFF2E7D32)
)
}
)
}
}
}
// Excluded Tags (RED)
if (excludedTags.isNotEmpty()) {
LazyRow(
horizontalArrangement = Arrangement.spacedBy(6.dp),
contentPadding = PaddingValues(vertical = 4.dp)
) {
items(excludedTags.toList()) { tag ->
FilterChipWithX(
label = tag,
color = Color(0xFFF44336).copy(alpha = 0.3f),
onTap = { searchViewModel.includeTag(tag) },
onRemove = { searchViewModel.removeTagFilter(tag) },
leadingIcon = {
Icon(
Icons.Default.LabelOff,
contentDescription = null,
modifier = Modifier.size(16.dp),
tint = Color(0xFFC62828)
)
}
)
}
}
}
}
}
}
// Results
when {
images.isEmpty() && searchViewModel.hasActiveFilters() -> NoResultsState()
images.isEmpty() && !searchViewModel.hasActiveFilters() -> EmptyState()
else -> {
LazyVerticalGrid(
columns = GridCells.Adaptive(minSize = 120.dp),
contentPadding = PaddingValues(16.dp),
horizontalArrangement = Arrangement.spacedBy(4.dp),
verticalArrangement = Arrangement.spacedBy(4.dp)
) {
items(images.size) { index ->
val imageWithTags = images[index]
Card(
modifier = Modifier
.aspectRatio(1f)
.clickable { onImageClick(imageWithTags.image.imageUri) },
shape = RoundedCornerShape(8.dp)
) {
AsyncImage(
model = imageWithTags.image.imageUri,
contentDescription = null,
modifier = Modifier.fillMaxSize()
)
}
}
}
}
}
}
// Face filter menu
if (showFaceFilterMenu) {
FaceFilterMenu(
currentFilter = faceFilter,
onSelect = { filter ->
searchViewModel.setFaceFilter(filter)
showFaceFilterMenu = false
},
onDismiss = { showFaceFilterMenu = false }
)
}
// People picker dialog
if (showPeoplePicker) {
PeoplePickerDialog(
people = availablePeople,
includedPeople = includedPeople,
excludedPeople = excludedPeople,
onInclude = { searchViewModel.includePerson(it) },
onExclude = { searchViewModel.excludePerson(it) },
onDismiss = { showPeoplePicker = false }
)
}
// Tag picker dialog
if (showTagPicker) {
TagPickerDialog(
tags = availableTags,
includedTags = includedTags,
excludedTags = excludedTags,
onInclude = { searchViewModel.includeTag(it) },
onExclude = { searchViewModel.excludeTag(it) },
onDismiss = { showTagPicker = false }
)
}
}
/**
* NEW: Filter chip with X button for easy removal
*/
@Composable
private fun FilterChipWithX(
label: String,
color: Color,
onTap: () -> Unit,
onRemove: () -> Unit,
leadingIcon: @Composable (() -> Unit)? = null
) {
Surface(
color = color,
shape = RoundedCornerShape(16.dp),
modifier = Modifier.height(32.dp)
) {
Row(
modifier = Modifier.padding(start = 8.dp, end = 4.dp),
verticalAlignment = Alignment.CenterVertically,
horizontalArrangement = Arrangement.spacedBy(6.dp)
) {
if (leadingIcon != null) {
leadingIcon()
}
Text(
text = label,
style = MaterialTheme.typography.labelMedium,
fontWeight = FontWeight.SemiBold,
modifier = Modifier.clickable(onClick = onTap)
)
IconButton(
onClick = onRemove,
modifier = Modifier.size(24.dp)
) {
Icon(
Icons.Default.Close,
contentDescription = "Remove",
modifier = Modifier.size(16.dp)
)
}
}
}
}
/**
* NEW: Face filter menu
*/
@Composable
private fun FaceFilterMenu(
currentFilter: FaceFilter,
onSelect: (FaceFilter) -> Unit,
onDismiss: () -> Unit
) {
AlertDialog(
onDismissRequest = onDismiss,
title = { Text("Filter by Faces") },
text = {
Column(verticalArrangement = Arrangement.spacedBy(8.dp)) {
FaceFilter.values().forEach { filter ->
Card(
modifier = Modifier
.fillMaxWidth()
.clickable { onSelect(filter) },
colors = CardDefaults.cardColors(
containerColor = if (filter == currentFilter) {
MaterialTheme.colorScheme.primaryContainer
} else {
MaterialTheme.colorScheme.surfaceVariant
}
)
) {
Row(
modifier = Modifier.padding(16.dp),
horizontalArrangement = Arrangement.spacedBy(12.dp),
verticalAlignment = Alignment.CenterVertically
) {
Icon(
when (filter) {
FaceFilter.ALL -> Icons.Default.FilterAlt
FaceFilter.HAS_FACES -> Icons.Default.Face
FaceFilter.NO_FACES -> Icons.Default.HideImage
},
contentDescription = null
)
Column {
Text(
filter.displayName,
style = MaterialTheme.typography.titleMedium,
fontWeight = FontWeight.Bold
)
Text(
when (filter) {
FaceFilter.ALL -> "Show all photos"
FaceFilter.HAS_FACES -> "Only photos with detected faces"
FaceFilter.NO_FACES -> "Only photos without faces"
},
style = MaterialTheme.typography.bodySmall,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
}
}
}
}
}
},
confirmButton = {
TextButton(onClick = onDismiss) {
Text("Done")
}
}
)
}
// ... Rest of dialogs remain the same ...
@Composable
private fun PeoplePickerDialog(
people: List<PersonEntity>,
includedPeople: Set<String>,
excludedPeople: Set<String>,
onInclude: (String) -> Unit,
onExclude: (String) -> Unit,
onDismiss: () -> Unit
) {
AlertDialog(
onDismissRequest = onDismiss,
title = { Text("Add Person Filter") },
text = {
Column(
modifier = Modifier
.fillMaxWidth()
.height(400.dp)
.verticalScroll(rememberScrollState()),
verticalArrangement = Arrangement.spacedBy(8.dp)
) {
Text(
"Tap to INCLUDE (green) • Long press to EXCLUDE (red)",
style = MaterialTheme.typography.bodySmall,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
people.forEach { person ->
val isIncluded = person.id in includedPeople
val isExcluded = person.id in excludedPeople
Card(
modifier = Modifier
.fillMaxWidth()
.clickable { onInclude(person.id) },
colors = CardDefaults.cardColors(
containerColor = when {
isIncluded -> Color(0xFF4CAF50).copy(alpha = 0.3f)
isExcluded -> Color(0xFFF44336).copy(alpha = 0.3f)
else -> MaterialTheme.colorScheme.surfaceVariant
}
)
) {
Row(
modifier = Modifier.padding(12.dp),
horizontalArrangement = Arrangement.SpaceBetween,
verticalAlignment = Alignment.CenterVertically
) {
Text(person.name, fontWeight = FontWeight.Medium)
Row(horizontalArrangement = Arrangement.spacedBy(4.dp)) {
IconButton(
onClick = { onInclude(person.id) },
colors = IconButtonDefaults.iconButtonColors(
containerColor = if (isIncluded) Color(0xFF4CAF50) else Color.Transparent
)
) {
Icon(Icons.Default.Check, "Include", tint = if (isIncluded) Color.White else MaterialTheme.colorScheme.onSurface)
}
IconButton(
onClick = { onExclude(person.id) },
colors = IconButtonDefaults.iconButtonColors(
containerColor = if (isExcluded) Color(0xFFF44336) else Color.Transparent
)
) {
Icon(Icons.Default.Close, "Exclude", tint = if (isExcluded) Color.White else MaterialTheme.colorScheme.onSurface)
}
}
}
}
}
}
},
confirmButton = {
TextButton(onClick = onDismiss) {
Text("Done")
}
}
)
}
@Composable
private fun TagPickerDialog(
tags: List<String>,
includedTags: Set<String>,
excludedTags: Set<String>,
onInclude: (String) -> Unit,
onExclude: (String) -> Unit,
onDismiss: () -> Unit
) {
AlertDialog(
onDismissRequest = onDismiss,
title = { Text("Add Tag Filter") },
text = {
Column(
modifier = Modifier
.fillMaxWidth()
.height(400.dp)
.verticalScroll(rememberScrollState()),
verticalArrangement = Arrangement.spacedBy(8.dp)
) {
Text(
"Tap to INCLUDE (green) • Long press to EXCLUDE (red)",
style = MaterialTheme.typography.bodySmall,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
tags.forEach { tagValue ->
val isIncluded = tagValue in includedTags
val isExcluded = tagValue in excludedTags
Card(
modifier = Modifier
.fillMaxWidth()
.clickable { onInclude(tagValue) },
colors = CardDefaults.cardColors(
containerColor = when {
isIncluded -> Color(0xFF4CAF50).copy(alpha = 0.3f)
isExcluded -> Color(0xFFF44336).copy(alpha = 0.3f)
else -> MaterialTheme.colorScheme.surfaceVariant
}
)
) {
Row(
modifier = Modifier.padding(12.dp),
horizontalArrangement = Arrangement.SpaceBetween,
verticalAlignment = Alignment.CenterVertically
) {
Text(tagValue, fontWeight = FontWeight.Medium)
Row(horizontalArrangement = Arrangement.spacedBy(4.dp)) {
IconButton(
onClick = { onInclude(tagValue) },
colors = IconButtonDefaults.iconButtonColors(
containerColor = if (isIncluded) Color(0xFF4CAF50) else Color.Transparent
)
) {
Icon(Icons.Default.Check, "Include", tint = if (isIncluded) Color.White else MaterialTheme.colorScheme.onSurface)
}
IconButton(
onClick = { onExclude(tagValue) },
colors = IconButtonDefaults.iconButtonColors(
containerColor = if (isExcluded) Color(0xFFF44336) else Color.Transparent
)
) {
Icon(Icons.Default.Close, "Exclude", tint = if (isExcluded) Color.White else MaterialTheme.colorScheme.onSurface)
}
}
}
}
}
}
},
confirmButton = {
TextButton(onClick = onDismiss) {
Text("Done")
}
}
)
}
@Composable
private fun EmptyState() {
Box(
modifier = Modifier
.fillMaxSize()
.padding(32.dp),
contentAlignment = Alignment.Center
) {
Column(
horizontalAlignment = Alignment.CenterHorizontally,
verticalArrangement = Arrangement.spacedBy(16.dp)
) {
Icon(
Icons.Default.Search,
contentDescription = null,
modifier = Modifier.size(64.dp),
tint = MaterialTheme.colorScheme.onSurfaceVariant.copy(alpha = 0.6f)
)
Text(
"Advanced Search",
style = MaterialTheme.typography.titleLarge,
fontWeight = FontWeight.Bold
)
Text(
"Add people, tags, or face filters to search",
style = MaterialTheme.typography.bodyMedium,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
}
}
}
@Composable
private fun NoResultsState() {
Box(
modifier = Modifier
.fillMaxSize()
.padding(32.dp),
contentAlignment = Alignment.Center
) {
Column(
horizontalAlignment = Alignment.CenterHorizontally,
verticalArrangement = Arrangement.spacedBy(12.dp)
) {
Icon(
Icons.Default.SearchOff,
contentDescription = null,
modifier = Modifier.size(64.dp),
tint = MaterialTheme.colorScheme.onSurfaceVariant.copy(alpha = 0.6f)
)
Text(
"No photos found",
style = MaterialTheme.typography.titleLarge,
fontWeight = FontWeight.Bold
)
Text(
"Try different filters",
style = MaterialTheme.typography.bodyMedium,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
}
}
}

View File

@@ -0,0 +1,340 @@
package com.placeholder.sherpai2.ui.search
import androidx.lifecycle.ViewModel
import androidx.lifecycle.viewModelScope
import com.placeholder.sherpai2.data.local.dao.ImageAggregateDao
import com.placeholder.sherpai2.data.local.dao.PersonDao
import com.placeholder.sherpai2.data.local.dao.TagDao
import com.placeholder.sherpai2.data.local.entity.ImageEntity
import com.placeholder.sherpai2.data.local.entity.PersonEntity
import com.placeholder.sherpai2.data.local.entity.PhotoFaceTagEntity
import com.placeholder.sherpai2.data.repository.FaceRecognitionRepository
import dagger.hilt.android.lifecycle.HiltViewModel
import kotlinx.coroutines.ExperimentalCoroutinesApi
import kotlinx.coroutines.flow.*
import kotlinx.coroutines.launch
import java.util.Calendar
import javax.inject.Inject
@OptIn(ExperimentalCoroutinesApi::class)
@HiltViewModel
class SearchViewModel @Inject constructor(
private val imageAggregateDao: ImageAggregateDao,
private val faceRecognitionRepository: FaceRecognitionRepository,
private val personDao: PersonDao,
private val tagDao: TagDao
) : ViewModel() {
private val _searchQuery = MutableStateFlow("")
val searchQuery: StateFlow<String> = _searchQuery.asStateFlow()
private val _includedPeople = MutableStateFlow<Set<String>>(emptySet())
val includedPeople: StateFlow<Set<String>> = _includedPeople.asStateFlow()
private val _excludedPeople = MutableStateFlow<Set<String>>(emptySet())
val excludedPeople: StateFlow<Set<String>> = _excludedPeople.asStateFlow()
private val _includedTags = MutableStateFlow<Set<String>>(emptySet())
val includedTags: StateFlow<Set<String>> = _includedTags.asStateFlow()
private val _excludedTags = MutableStateFlow<Set<String>>(emptySet())
val excludedTags: StateFlow<Set<String>> = _excludedTags.asStateFlow()
private val _dateRange = MutableStateFlow(DateRange.ALL_TIME)
val dateRange: StateFlow<DateRange> = _dateRange.asStateFlow()
private val _faceFilter = MutableStateFlow(FaceFilter.ALL)
val faceFilter: StateFlow<FaceFilter> = _faceFilter.asStateFlow()
private val _availablePeople = MutableStateFlow<List<PersonEntity>>(emptyList())
val availablePeople: StateFlow<List<PersonEntity>> = _availablePeople.asStateFlow()
private val _availableTags = MutableStateFlow<List<String>>(emptyList())
val availableTags: StateFlow<List<String>> = _availableTags.asStateFlow()
private val personCache = mutableMapOf<String, String>()
init {
loadAvailableFilters()
buildPersonCache()
}
private fun buildPersonCache() {
viewModelScope.launch {
val people = personDao.getAllPersons()
people.forEach { person ->
val stats = faceRecognitionRepository.getPersonFaceStats(person.id)
if (stats != null) {
personCache[stats.faceModelId] = person.id
}
}
}
}
fun searchImages(): Flow<List<ImageWithFaceTags>> {
return combine(
_searchQuery,
_includedPeople,
_excludedPeople,
_includedTags,
_excludedTags,
_dateRange,
_faceFilter
) { values: Array<*> ->
@Suppress("UNCHECKED_CAST")
SearchCriteria(
query = values[0] as String,
includedPeople = values[1] as Set<String>,
excludedPeople = values[2] as Set<String>,
includedTags = values[3] as Set<String>,
excludedTags = values[4] as Set<String>,
dateRange = values[5] as DateRange,
faceFilter = values[6] as FaceFilter
)
}.flatMapLatest { criteria ->
imageAggregateDao.observeAllImagesWithEverything()
.map { imagesList ->
imagesList.mapNotNull { imageWithEverything ->
// Apply date filter
if (!isInDateRange(imageWithEverything.image.capturedAt, criteria.dateRange)) {
return@mapNotNull null
}
// Apply face filter - ONLY when cache is explicitly set
when (criteria.faceFilter) {
FaceFilter.HAS_FACES -> {
// Only show images where hasFaces is EXPLICITLY true
if (imageWithEverything.image.hasFaces != true) {
return@mapNotNull null
}
}
FaceFilter.NO_FACES -> {
// Only show images where hasFaces is EXPLICITLY false
if (imageWithEverything.image.hasFaces != false) {
return@mapNotNull null
}
}
FaceFilter.ALL -> {
// Show all images (null, true, or false)
}
}
val personIds = imageWithEverything.faceTags
.mapNotNull { faceTag -> personCache[faceTag.faceModelId] }
.toSet()
val imageTags = imageWithEverything.tags
.map { it.value }
.toSet()
val passesFilter = applyBooleanLogic(
personIds = personIds,
imageTags = imageTags,
criteria = criteria
)
if (passesFilter) {
val persons = personIds.mapNotNull { personId ->
_availablePeople.value.find { it.id == personId }
}
ImageWithFaceTags(
image = imageWithEverything.image,
faceTags = imageWithEverything.faceTags,
persons = persons
)
} else {
null
}
}.sortedByDescending { it.image.capturedAt }
}
}
}
private fun applyBooleanLogic(
personIds: Set<String>,
imageTags: Set<String>,
criteria: SearchCriteria
): Boolean {
val hasAllIncludedPeople = if (criteria.includedPeople.isNotEmpty()) {
criteria.includedPeople.all { it in personIds }
} else true
val hasNoExcludedPeople = if (criteria.excludedPeople.isNotEmpty()) {
criteria.excludedPeople.none { it in personIds }
} else true
val hasAllIncludedTags = if (criteria.includedTags.isNotEmpty()) {
criteria.includedTags.all { it in imageTags }
} else true
val hasNoExcludedTags = if (criteria.excludedTags.isNotEmpty()) {
criteria.excludedTags.none { it in imageTags }
} else true
val matchesTextSearch = if (criteria.query.isNotBlank()) {
val normalizedQuery = criteria.query.trim().lowercase()
imageTags.any { tag -> tag.lowercase().contains(normalizedQuery) }
} else true
return hasAllIncludedPeople && hasNoExcludedPeople &&
hasAllIncludedTags && hasNoExcludedTags &&
matchesTextSearch
}
private fun loadAvailableFilters() {
viewModelScope.launch {
val people = personDao.getAllPersons()
_availablePeople.value = people.sortedBy { it.name }
val tags = tagDao.getByType("SYSTEM")
val tagsWithUsage = tags.map { tag ->
tag to tagDao.getTagUsageCount(tag.tagId)
}
_availableTags.value = tagsWithUsage
.sortedByDescending { (_, usageCount) -> usageCount }
.take(30)
.map { (tag, _) -> tag.value }
}
}
fun includePerson(personId: String) {
_includedPeople.value = _includedPeople.value + personId
_excludedPeople.value = _excludedPeople.value - personId
}
fun excludePerson(personId: String) {
_excludedPeople.value = _excludedPeople.value + personId
_includedPeople.value = _includedPeople.value - personId
}
fun removePersonFilter(personId: String) {
_includedPeople.value = _includedPeople.value - personId
_excludedPeople.value = _excludedPeople.value - personId
}
fun includeTag(tagValue: String) {
_includedTags.value = _includedTags.value + tagValue
_excludedTags.value = _excludedTags.value - tagValue
}
fun excludeTag(tagValue: String) {
_excludedTags.value = _excludedTags.value + tagValue
_includedTags.value = _includedTags.value - tagValue
}
fun removeTagFilter(tagValue: String) {
_includedTags.value = _includedTags.value - tagValue
_excludedTags.value = _excludedTags.value - tagValue
}
fun setSearchQuery(query: String) {
_searchQuery.value = query
}
fun setDateRange(range: DateRange) {
_dateRange.value = range
}
fun setFaceFilter(filter: FaceFilter) {
_faceFilter.value = filter
}
fun clearAllFilters() {
_searchQuery.value = ""
_includedPeople.value = emptySet()
_excludedPeople.value = emptySet()
_includedTags.value = emptySet()
_excludedTags.value = emptySet()
_dateRange.value = DateRange.ALL_TIME
_faceFilter.value = FaceFilter.ALL
}
fun hasActiveFilters(): Boolean {
return _searchQuery.value.isNotBlank() ||
_includedPeople.value.isNotEmpty() ||
_excludedPeople.value.isNotEmpty() ||
_includedTags.value.isNotEmpty() ||
_excludedTags.value.isNotEmpty() ||
_dateRange.value != DateRange.ALL_TIME ||
_faceFilter.value != FaceFilter.ALL
}
fun getSearchSummary(): String {
val parts = mutableListOf<String>()
if (_includedPeople.value.isNotEmpty()) parts.add("WITH: ${_includedPeople.value.size} people")
if (_excludedPeople.value.isNotEmpty()) parts.add("WITHOUT: ${_excludedPeople.value.size} people")
if (_includedTags.value.isNotEmpty()) parts.add("HAS: ${_includedTags.value.size} tags")
if (_excludedTags.value.isNotEmpty()) parts.add("NOT: ${_excludedTags.value.size} tags")
if (_dateRange.value != DateRange.ALL_TIME) parts.add(_dateRange.value.displayName)
return parts.joinToString("")
}
private fun isInDateRange(timestamp: Long, range: DateRange): Boolean = when (range) {
DateRange.ALL_TIME -> true
DateRange.TODAY -> isToday(timestamp)
DateRange.THIS_WEEK -> isThisWeek(timestamp)
DateRange.THIS_MONTH -> isThisMonth(timestamp)
DateRange.THIS_YEAR -> isThisYear(timestamp)
}
private fun isToday(timestamp: Long): Boolean {
val today = Calendar.getInstance()
val date = Calendar.getInstance().apply { timeInMillis = timestamp }
return today.get(Calendar.YEAR) == date.get(Calendar.YEAR) &&
today.get(Calendar.DAY_OF_YEAR) == date.get(Calendar.DAY_OF_YEAR)
}
private fun isThisWeek(timestamp: Long): Boolean {
val today = Calendar.getInstance()
val date = Calendar.getInstance().apply { timeInMillis = timestamp }
return today.get(Calendar.YEAR) == date.get(Calendar.YEAR) &&
today.get(Calendar.WEEK_OF_YEAR) == date.get(Calendar.WEEK_OF_YEAR)
}
private fun isThisMonth(timestamp: Long): Boolean {
val today = Calendar.getInstance()
val date = Calendar.getInstance().apply { timeInMillis = timestamp }
return today.get(Calendar.YEAR) == date.get(Calendar.YEAR) &&
today.get(Calendar.MONTH) == date.get(Calendar.MONTH)
}
private fun isThisYear(timestamp: Long): Boolean {
val today = Calendar.getInstance()
val date = Calendar.getInstance().apply { timeInMillis = timestamp }
return today.get(Calendar.YEAR) == date.get(Calendar.YEAR)
}
}
private data class SearchCriteria(
val query: String,
val includedPeople: Set<String>,
val excludedPeople: Set<String>,
val includedTags: Set<String>,
val excludedTags: Set<String>,
val dateRange: DateRange,
val faceFilter: FaceFilter
)
data class ImageWithFaceTags(
val image: ImageEntity,
val faceTags: List<PhotoFaceTagEntity>,
val persons: List<PersonEntity>
)
enum class DateRange(val displayName: String) {
ALL_TIME("All Time"),
TODAY("Today"),
THIS_WEEK("This Week"),
THIS_MONTH("This Month"),
THIS_YEAR("This Year")
}
enum class FaceFilter(val displayName: String) {
ALL("All Photos"),
HAS_FACES("Has Faces"),
NO_FACES("No Faces")
}
@Deprecated("No longer used")
enum class DisplayMode { SIMPLE, VERBOSE }

View File

@@ -0,0 +1,30 @@
package com.placeholder.sherpai2.ui.search.components
import androidx.compose.foundation.Image
import androidx.compose.foundation.layout.aspectRatio
import androidx.compose.foundation.layout.fillMaxWidth
import androidx.compose.runtime.Composable
import androidx.compose.ui.Modifier
import coil.compose.rememberAsyncImagePainter
import com.placeholder.sherpai2.data.local.entity.ImageEntity
/**
* ImageGridItem
*
* Minimal thumbnail preview.
* No click handling yet.
*/
@Composable
fun ImageGridItem(
image: ImageEntity,
modifier: Modifier = Modifier,
onClick: (() -> Unit)? = null
) {
Image(
painter = rememberAsyncImagePainter(image.imageUri),
contentDescription = null,
modifier = Modifier
.fillMaxWidth()
.aspectRatio(1f)
)
}

View File

@@ -0,0 +1,638 @@
package com.placeholder.sherpai2.ui.tags
import androidx.compose.animation.AnimatedVisibility
import androidx.compose.animation.expandVertically
import androidx.compose.animation.fadeIn
import androidx.compose.animation.fadeOut
import androidx.compose.animation.shrinkVertically
import androidx.compose.foundation.background
import androidx.compose.foundation.clickable
import androidx.compose.foundation.layout.*
import androidx.compose.foundation.lazy.LazyColumn
import androidx.compose.foundation.lazy.items
import androidx.compose.foundation.shape.RoundedCornerShape
import androidx.compose.material.icons.Icons
import androidx.compose.material.icons.filled.*
import androidx.compose.material3.*
import androidx.compose.runtime.*
import androidx.compose.ui.Alignment
import androidx.compose.ui.Modifier
import androidx.compose.ui.graphics.Brush
import androidx.compose.ui.graphics.vector.ImageVector
import androidx.compose.ui.text.font.FontWeight
import androidx.compose.ui.unit.dp
import androidx.hilt.navigation.compose.hiltViewModel
import com.placeholder.sherpai2.data.local.entity.TagWithUsage
/**
* CLEANED TagManagementScreen - No Scaffold wrapper
*
* Removed:
* - Scaffold wrapper (line 38)
* - Moved FAB inline as part of content
*
* Features:
* - Tag list with usage counts
* - Search functionality
* - Scanning progress
* - Delete tags
* - System/User tag distinction
*/
@Composable
fun TagManagementScreen(
viewModel: TagManagementViewModel = hiltViewModel(),
modifier: Modifier = Modifier
) {
val uiState by viewModel.uiState.collectAsState()
val scanningState by viewModel.scanningState.collectAsState()
var showAddTagDialog by remember { mutableStateOf(false) }
var showScanMenu by remember { mutableStateOf(false) }
var searchQuery by remember { mutableStateOf("") }
Box(modifier = modifier.fillMaxSize()) {
Column(modifier = Modifier.fillMaxSize()) {
// Stats Bar
StatsBar(uiState)
// Search Bar
SearchBar(
searchQuery = searchQuery,
onSearchChange = {
searchQuery = it
viewModel.searchTags(it)
}
)
// Scanning Progress
AnimatedVisibility(
visible = scanningState !is TagManagementViewModel.TagScanningState.Idle,
enter = expandVertically() + fadeIn(),
exit = shrinkVertically() + fadeOut()
) {
ScanningProgress(scanningState, viewModel)
}
// Tag List
when (val state = uiState) {
is TagManagementViewModel.TagUiState.Loading -> {
Box(
modifier = Modifier.fillMaxSize(),
contentAlignment = Alignment.Center
) {
CircularProgressIndicator()
}
}
is TagManagementViewModel.TagUiState.Success -> {
if (state.tags.isEmpty()) {
EmptyTagsView()
} else {
TagList(
tags = state.tags,
onDeleteTag = { viewModel.deleteTag(it) }
)
}
}
is TagManagementViewModel.TagUiState.Error -> {
Box(
modifier = Modifier.fillMaxSize(),
contentAlignment = Alignment.Center
) {
Column(
horizontalAlignment = Alignment.CenterHorizontally,
verticalArrangement = Arrangement.spacedBy(8.dp)
) {
Icon(
Icons.Default.Error,
contentDescription = null,
modifier = Modifier.size(48.dp),
tint = MaterialTheme.colorScheme.error
)
Text(
text = state.message,
color = MaterialTheme.colorScheme.error
)
}
}
}
}
}
// FAB (inline, positioned over content)
ScanFAB(
showMenu = showScanMenu,
onToggleMenu = { showScanMenu = !showScanMenu },
onScanAll = {
viewModel.scanForAllTags()
showScanMenu = false
},
onScanBase = {
viewModel.scanForBaseTags()
showScanMenu = false
},
onScanRelationships = {
viewModel.scanForRelationshipTags()
showScanMenu = false
},
modifier = Modifier
.align(Alignment.BottomEnd)
.padding(16.dp)
)
}
// Add Tag Dialog
if (showAddTagDialog) {
AddTagDialog(
onDismiss = { showAddTagDialog = false },
onConfirm = { tagName ->
viewModel.createUserTag(tagName)
showAddTagDialog = false
}
)
}
}
/**
* Stats bar at top
*/
@Composable
private fun StatsBar(uiState: TagManagementViewModel.TagUiState) {
val (totalTags, totalPhotos) = when (uiState) {
is TagManagementViewModel.TagUiState.Success -> {
val photoCount: Int = uiState.tags.sumOf { it.usageCount }
uiState.tags.size to photoCount
}
else -> 0 to 0
}
Surface(
modifier = Modifier.fillMaxWidth(),
color = MaterialTheme.colorScheme.primaryContainer.copy(alpha = 0.3f)
) {
Row(
modifier = Modifier
.fillMaxWidth()
.padding(16.dp),
horizontalArrangement = Arrangement.SpaceEvenly
) {
StatItem(
icon = Icons.Default.Label,
value = totalTags.toString(),
label = "Tags"
)
VerticalDivider(
modifier = Modifier.height(48.dp),
color = MaterialTheme.colorScheme.outline.copy(alpha = 0.3f)
)
StatItem(
icon = Icons.Default.Photo,
value = totalPhotos.toString(),
label = "Tagged Photos"
)
}
}
}
@Composable
private fun StatItem(icon: ImageVector, value: String, label: String) {
Column(
horizontalAlignment = Alignment.CenterHorizontally,
verticalArrangement = Arrangement.spacedBy(4.dp)
) {
Icon(
icon,
contentDescription = null,
modifier = Modifier.size(24.dp),
tint = MaterialTheme.colorScheme.primary
)
Text(
value,
style = MaterialTheme.typography.headlineSmall,
fontWeight = FontWeight.Bold
)
Text(
label,
style = MaterialTheme.typography.labelMedium,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
}
}
/**
* Search bar
*/
@Composable
private fun SearchBar(
searchQuery: String,
onSearchChange: (String) -> Unit
) {
OutlinedTextField(
value = searchQuery,
onValueChange = onSearchChange,
modifier = Modifier
.fillMaxWidth()
.padding(16.dp),
placeholder = { Text("Search tags...") },
leadingIcon = { Icon(Icons.Default.Search, contentDescription = null) },
trailingIcon = {
if (searchQuery.isNotEmpty()) {
IconButton(onClick = { onSearchChange("") }) {
Icon(Icons.Default.Clear, "Clear")
}
}
},
singleLine = true,
shape = RoundedCornerShape(16.dp)
)
}
/**
* Scanning progress indicator
*/
@Composable
private fun ScanningProgress(
scanningState: TagManagementViewModel.TagScanningState,
viewModel: TagManagementViewModel
) {
when (scanningState) {
is TagManagementViewModel.TagScanningState.Scanning -> {
Card(
modifier = Modifier
.fillMaxWidth()
.padding(horizontal = 16.dp, vertical = 8.dp),
colors = CardDefaults.cardColors(
containerColor = MaterialTheme.colorScheme.primaryContainer.copy(alpha = 0.5f)
)
) {
Column(
modifier = Modifier.padding(16.dp),
verticalArrangement = Arrangement.spacedBy(8.dp)
) {
Row(
modifier = Modifier.fillMaxWidth(),
horizontalArrangement = Arrangement.SpaceBetween,
verticalAlignment = Alignment.CenterVertically
) {
Text(
"Scanning: ${scanningState.scanType}",
style = MaterialTheme.typography.titleSmall,
fontWeight = FontWeight.SemiBold
)
Text(
"${scanningState.progress}/${scanningState.total}",
style = MaterialTheme.typography.bodySmall
)
}
LinearProgressIndicator(
progress = {
if (scanningState.total > 0) {
scanningState.progress.toFloat() / scanningState.total.toFloat()
} else {
0f
}
},
modifier = Modifier.fillMaxWidth()
)
Text(
"Tags applied: ${scanningState.tagsApplied}",
style = MaterialTheme.typography.bodySmall,
color = MaterialTheme.colorScheme.primary
)
if (scanningState.currentImage.isNotEmpty()) {
Text(
"Current: ${scanningState.currentImage}",
style = MaterialTheme.typography.labelSmall,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
}
}
}
}
is TagManagementViewModel.TagScanningState.Complete -> {
Card(
modifier = Modifier
.fillMaxWidth()
.padding(horizontal = 16.dp, vertical = 8.dp),
colors = CardDefaults.cardColors(
containerColor = MaterialTheme.colorScheme.primaryContainer
)
) {
Row(
modifier = Modifier.padding(16.dp),
horizontalArrangement = Arrangement.spacedBy(12.dp),
verticalAlignment = Alignment.CenterVertically
) {
Icon(
Icons.Default.CheckCircle,
contentDescription = null,
tint = MaterialTheme.colorScheme.primary
)
Column {
Text(
"Scan Complete!",
style = MaterialTheme.typography.titleSmall,
fontWeight = FontWeight.Bold
)
Text(
"Processed: ${scanningState.imagesProcessed} images",
style = MaterialTheme.typography.bodySmall
)
Text(
"Applied: ${scanningState.tagsApplied} tags",
style = MaterialTheme.typography.bodySmall
)
if (scanningState.newTagsCreated > 0) {
Text(
"Created: ${scanningState.newTagsCreated} new tags",
style = MaterialTheme.typography.bodySmall,
color = MaterialTheme.colorScheme.primary
)
}
}
}
}
}
else -> {}
}
}
/**
* Tag list
*/
@Composable
private fun TagList(
tags: List<TagWithUsage>,
onDeleteTag: (String) -> Unit
) {
LazyColumn(
modifier = Modifier.fillMaxSize(),
contentPadding = PaddingValues(16.dp),
verticalArrangement = Arrangement.spacedBy(8.dp)
) {
items(tags) { tagWithUsage ->
TagCard(
tagWithUsage = tagWithUsage,
onDelete = { onDeleteTag(tagWithUsage.tagId) }
)
}
}
}
/**
* Individual tag card
*/
@Composable
private fun TagCard(
tagWithUsage: TagWithUsage,
onDelete: () -> Unit
) {
val isSystemTag = tagWithUsage.type == "SYSTEM"
Card(
modifier = Modifier.fillMaxWidth(),
elevation = CardDefaults.cardElevation(defaultElevation = 2.dp)
) {
Row(
modifier = Modifier.padding(16.dp),
horizontalArrangement = Arrangement.SpaceBetween,
verticalAlignment = Alignment.CenterVertically
) {
Row(
modifier = Modifier.weight(1f),
horizontalArrangement = Arrangement.spacedBy(12.dp),
verticalAlignment = Alignment.CenterVertically
) {
// Tag icon
Surface(
modifier = Modifier.size(40.dp),
shape = RoundedCornerShape(8.dp),
color = if (isSystemTag)
MaterialTheme.colorScheme.primaryContainer
else
MaterialTheme.colorScheme.secondaryContainer
) {
Box(contentAlignment = Alignment.Center) {
Icon(
if (isSystemTag) Icons.Default.AutoAwesome else Icons.Default.Label,
contentDescription = null,
modifier = Modifier.size(20.dp),
tint = if (isSystemTag)
MaterialTheme.colorScheme.onPrimaryContainer
else
MaterialTheme.colorScheme.onSecondaryContainer
)
}
}
// Tag info
Column {
Row(
horizontalArrangement = Arrangement.spacedBy(8.dp),
verticalAlignment = Alignment.CenterVertically
) {
Text(
text = tagWithUsage.value,
style = MaterialTheme.typography.titleMedium,
fontWeight = FontWeight.SemiBold
)
if (isSystemTag) {
Surface(
shape = RoundedCornerShape(4.dp),
color = MaterialTheme.colorScheme.primary.copy(alpha = 0.1f)
) {
Text(
"SYSTEM",
modifier = Modifier.padding(horizontal = 6.dp, vertical = 2.dp),
style = MaterialTheme.typography.labelSmall,
color = MaterialTheme.colorScheme.primary
)
}
}
}
Text(
text = "${tagWithUsage.usageCount} ${if (tagWithUsage.usageCount == 1) "photo" else "photos"}",
style = MaterialTheme.typography.bodySmall,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
}
}
// Delete button (only for user tags)
if (!isSystemTag) {
IconButton(onClick = onDelete) {
Icon(
Icons.Default.Delete,
contentDescription = "Delete",
tint = MaterialTheme.colorScheme.error
)
}
}
}
}
}
/**
* Empty state
*/
@Composable
private fun EmptyTagsView() {
Box(
modifier = Modifier
.fillMaxSize()
.padding(32.dp),
contentAlignment = Alignment.Center
) {
Column(
horizontalAlignment = Alignment.CenterHorizontally,
verticalArrangement = Arrangement.spacedBy(16.dp)
) {
Icon(
Icons.Default.LabelOff,
contentDescription = null,
modifier = Modifier.size(64.dp),
tint = MaterialTheme.colorScheme.onSurfaceVariant.copy(alpha = 0.6f)
)
Text(
"No Tags Yet",
style = MaterialTheme.typography.titleLarge,
fontWeight = FontWeight.Bold
)
Text(
"Scan your photos to generate tags automatically",
style = MaterialTheme.typography.bodyMedium,
color = MaterialTheme.colorScheme.onSurfaceVariant,
textAlign = androidx.compose.ui.text.style.TextAlign.Center
)
}
}
}
/**
* Floating Action Button with scan menu
*/
@Composable
private fun ScanFAB(
showMenu: Boolean,
onToggleMenu: () -> Unit,
onScanAll: () -> Unit,
onScanBase: () -> Unit,
onScanRelationships: () -> Unit,
modifier: Modifier = Modifier
) {
Column(
modifier = modifier,
horizontalAlignment = Alignment.End,
verticalArrangement = Arrangement.spacedBy(8.dp)
) {
// Menu options
AnimatedVisibility(visible = showMenu) {
Column(
horizontalAlignment = Alignment.End,
verticalArrangement = Arrangement.spacedBy(8.dp)
) {
SmallFAB(
icon = Icons.Default.AutoFixHigh,
text = "Scan All",
onClick = onScanAll
)
SmallFAB(
icon = Icons.Default.PhotoCamera,
text = "Base Tags",
onClick = onScanBase
)
SmallFAB(
icon = Icons.Default.People,
text = "Relationships",
onClick = onScanRelationships
)
}
}
// Main FAB
ExtendedFloatingActionButton(
onClick = onToggleMenu,
icon = {
Icon(
if (showMenu) Icons.Default.Close else Icons.Default.AutoFixHigh,
"Scan"
)
},
text = { Text(if (showMenu) "Close" else "Scan Tags") },
containerColor = MaterialTheme.colorScheme.primaryContainer,
contentColor = MaterialTheme.colorScheme.onPrimaryContainer
)
}
}
@Composable
private fun SmallFAB(
icon: ImageVector,
text: String,
onClick: () -> Unit
) {
Row(
verticalAlignment = Alignment.CenterVertically,
horizontalArrangement = Arrangement.spacedBy(8.dp)
) {
Surface(
shape = RoundedCornerShape(8.dp),
color = MaterialTheme.colorScheme.surface,
shadowElevation = 2.dp
) {
Text(
text,
modifier = Modifier.padding(horizontal = 12.dp, vertical = 8.dp),
style = MaterialTheme.typography.bodySmall,
fontWeight = FontWeight.Medium
)
}
FloatingActionButton(
onClick = onClick,
modifier = Modifier.size(48.dp),
containerColor = MaterialTheme.colorScheme.secondaryContainer,
contentColor = MaterialTheme.colorScheme.onSecondaryContainer
) {
Icon(icon, contentDescription = text, modifier = Modifier.size(20.dp))
}
}
}
/**
* Add tag dialog
*/
@Composable
private fun AddTagDialog(
onDismiss: () -> Unit,
onConfirm: (String) -> Unit
) {
var tagName by remember { mutableStateOf("") }
AlertDialog(
onDismissRequest = onDismiss,
icon = { Icon(Icons.Default.Add, contentDescription = null) },
title = { Text("Add Custom Tag") },
text = {
OutlinedTextField(
value = tagName,
onValueChange = { tagName = it },
label = { Text("Tag Name") },
singleLine = true,
modifier = Modifier.fillMaxWidth()
)
},
confirmButton = {
Button(
onClick = { onConfirm(tagName) },
enabled = tagName.isNotBlank()
) {
Text("Add")
}
},
dismissButton = {
TextButton(onClick = onDismiss) {
Text("Cancel")
}
}
)
}

View File

@@ -0,0 +1,398 @@
package com.placeholder.sherpai2.ui.tags
import android.app.Application
import android.graphics.BitmapFactory
import android.net.Uri
import androidx.lifecycle.AndroidViewModel
import androidx.lifecycle.viewModelScope
import com.google.mlkit.vision.common.InputImage
import com.google.mlkit.vision.face.FaceDetection
import com.google.mlkit.vision.face.FaceDetectorOptions
import com.placeholder.sherpai2.data.local.dao.ImageTagDao
import com.placeholder.sherpai2.data.local.dao.TagDao
import com.placeholder.sherpai2.data.local.entity.TagEntity
import com.placeholder.sherpai2.data.local.entity.TagWithUsage
import com.placeholder.sherpai2.data.repository.DetectedFace
import com.placeholder.sherpai2.data.service.AutoTaggingService
import com.placeholder.sherpai2.domain.repository.ImageRepository
import com.placeholder.sherpai2.util.DiagnosticLogger
import dagger.hilt.android.lifecycle.HiltViewModel
import kotlinx.coroutines.Dispatchers
import kotlinx.coroutines.flow.MutableStateFlow
import kotlinx.coroutines.flow.StateFlow
import kotlinx.coroutines.flow.asStateFlow
import kotlinx.coroutines.flow.first
import kotlinx.coroutines.launch
import kotlinx.coroutines.withContext
import kotlinx.coroutines.tasks.await
import javax.inject.Inject
@HiltViewModel
class TagManagementViewModel @Inject constructor(
application: Application,
private val tagDao: TagDao,
private val imageTagDao: ImageTagDao,
private val imageRepository: ImageRepository,
private val autoTaggingService: AutoTaggingService
) : AndroidViewModel(application) {
private val _uiState = MutableStateFlow<TagUiState>(TagUiState.Loading)
val uiState: StateFlow<TagUiState> = _uiState.asStateFlow()
private val _scanningState = MutableStateFlow<TagScanningState>(TagScanningState.Idle)
val scanningState: StateFlow<TagScanningState> = _scanningState.asStateFlow()
private val faceDetector by lazy {
val options = FaceDetectorOptions.Builder()
.setPerformanceMode(FaceDetectorOptions.PERFORMANCE_MODE_ACCURATE)
.setLandmarkMode(FaceDetectorOptions.LANDMARK_MODE_NONE)
.setClassificationMode(FaceDetectorOptions.CLASSIFICATION_MODE_NONE)
.setMinFaceSize(0.10f)
.build()
FaceDetection.getClient(options)
}
sealed class TagUiState {
object Loading : TagUiState()
data class Success(
val tags: List<TagWithUsage>,
val totalTags: Int,
val systemTags: Int,
val userTags: Int
) : TagUiState()
data class Error(val message: String) : TagUiState()
}
sealed class TagScanningState {
object Idle : TagScanningState()
data class Scanning(
val scanType: ScanType,
val progress: Int,
val total: Int,
val tagsApplied: Int,
val currentImage: String = ""
) : TagScanningState()
data class Complete(
val scanType: ScanType,
val imagesProcessed: Int,
val tagsApplied: Int,
val newTagsCreated: Int = 0
) : TagScanningState()
data class Error(val message: String) : TagScanningState()
}
enum class ScanType {
BASE_TAGS, // Face count, orientation, resolution, time-of-day
RELATIONSHIP_TAGS, // Family, friend, colleague from person entities
BIRTHDAY_TAGS, // Birthday tags for DOB matches
SCENE_TAGS, // Indoor/outdoor estimation
ALL // Run all scans
}
init {
loadTags()
}
fun loadTags() {
viewModelScope.launch {
try {
_uiState.value = TagUiState.Loading
val tagsWithUsage = tagDao.getMostUsedTags(1000) // Get all tags
val systemTags = tagsWithUsage.count { it.type == "SYSTEM" }
val userTags = tagsWithUsage.count { it.type == "GENERIC" }
_uiState.value = TagUiState.Success(
tags = tagsWithUsage,
totalTags = tagsWithUsage.size,
systemTags = systemTags,
userTags = userTags
)
} catch (e: Exception) {
_uiState.value = TagUiState.Error(
e.message ?: "Failed to load tags"
)
}
}
}
fun createUserTag(tagName: String) {
viewModelScope.launch {
try {
val trimmedName = tagName.trim().lowercase()
if (trimmedName.isEmpty()) {
_uiState.value = TagUiState.Error("Tag name cannot be empty")
return@launch
}
// Check if tag already exists
val existing = tagDao.getByValue(trimmedName)
if (existing != null) {
_uiState.value = TagUiState.Error("Tag '$trimmedName' already exists")
return@launch
}
val newTag = TagEntity.createUserTag(trimmedName)
tagDao.insert(newTag)
loadTags()
} catch (e: Exception) {
_uiState.value = TagUiState.Error(
"Failed to create tag: ${e.message}"
)
}
}
}
fun deleteTag(tagId: String) {
viewModelScope.launch {
try {
tagDao.delete(tagId)
loadTags()
} catch (e: Exception) {
_uiState.value = TagUiState.Error(
"Failed to delete tag: ${e.message}"
)
}
}
}
fun searchTags(query: String) {
viewModelScope.launch {
try {
val results = if (query.isBlank()) {
tagDao.getMostUsedTags(1000)
} else {
tagDao.searchTagsWithUsage(query, 100)
}
val systemTags = results.count { it.type == "SYSTEM" }
val userTags = results.count { it.type == "GENERIC" }
_uiState.value = TagUiState.Success(
tags = results,
totalTags = results.size,
systemTags = systemTags,
userTags = userTags
)
} catch (e: Exception) {
_uiState.value = TagUiState.Error("Search failed: ${e.message}")
}
}
}
// ======================
// AUTO-TAGGING SCANS
// ======================
/**
* Scan library for base tags (face count, orientation, time, quality, scene)
*/
fun scanForBaseTags() {
performScan(ScanType.BASE_TAGS)
}
/**
* Scan for relationship tags (family, friend, colleague)
*/
fun scanForRelationshipTags() {
performScan(ScanType.RELATIONSHIP_TAGS)
}
/**
* Scan for birthday tags
*/
fun scanForBirthdayTags() {
performScan(ScanType.BIRTHDAY_TAGS)
}
/**
* Scan for scene tags (indoor/outdoor)
*/
fun scanForSceneTags() {
performScan(ScanType.SCENE_TAGS)
}
/**
* Scan for ALL tags
*/
fun scanForAllTags() {
performScan(ScanType.ALL)
}
private fun performScan(scanType: ScanType) {
viewModelScope.launch {
try {
DiagnosticLogger.i("=== STARTING TAG SCAN: $scanType ===")
_scanningState.value = TagScanningState.Scanning(
scanType = scanType,
progress = 0,
total = 0,
tagsApplied = 0
)
val allImages = imageRepository.getAllImages().first()
var tagsApplied = 0
var newTagsCreated = 0
DiagnosticLogger.i("Processing ${allImages.size} images")
allImages.forEachIndexed { index, imageWithEverything ->
val image = imageWithEverything.image
_scanningState.value = TagScanningState.Scanning(
scanType = scanType,
progress = index + 1,
total = allImages.size,
tagsApplied = tagsApplied,
currentImage = image.imageId.take(8)
)
when (scanType) {
ScanType.BASE_TAGS -> {
tagsApplied += scanImageForBaseTags(image.imageUri, image)
}
ScanType.SCENE_TAGS -> {
tagsApplied += scanImageForSceneTags(image.imageUri, image)
}
ScanType.RELATIONSHIP_TAGS -> {
// Handled at person level, not per-image
}
ScanType.BIRTHDAY_TAGS -> {
// Handled at person level, not per-image
}
ScanType.ALL -> {
tagsApplied += scanImageForBaseTags(image.imageUri, image)
tagsApplied += scanImageForSceneTags(image.imageUri, image)
}
}
}
// Handle person-level scans
if (scanType == ScanType.RELATIONSHIP_TAGS || scanType == ScanType.ALL) {
DiagnosticLogger.i("Scanning relationship tags...")
tagsApplied += autoTaggingService.autoTagAllRelationships()
}
if (scanType == ScanType.BIRTHDAY_TAGS || scanType == ScanType.ALL) {
DiagnosticLogger.i("Scanning birthday tags...")
tagsApplied += autoTaggingService.autoTagAllBirthdays(daysRange = 3)
}
DiagnosticLogger.i("=== SCAN COMPLETE ===")
DiagnosticLogger.i("Images processed: ${allImages.size}")
DiagnosticLogger.i("Tags applied: $tagsApplied")
_scanningState.value = TagScanningState.Complete(
scanType = scanType,
imagesProcessed = allImages.size,
tagsApplied = tagsApplied,
newTagsCreated = newTagsCreated
)
loadTags()
} catch (e: Exception) {
DiagnosticLogger.e("Scan failed", e)
_scanningState.value = TagScanningState.Error(
"Scan failed: ${e.message}"
)
}
}
}
private suspend fun scanImageForBaseTags(
imageUri: String,
image: com.placeholder.sherpai2.data.local.entity.ImageEntity
): Int = withContext(Dispatchers.Default) {
try {
val uri = Uri.parse(imageUri)
val inputStream = getApplication<Application>().contentResolver.openInputStream(uri)
val bitmap = BitmapFactory.decodeStream(inputStream)
inputStream?.close()
if (bitmap == null) return@withContext 0
// Detect faces
val detectedFaces = detectFaces(bitmap)
// Auto-tag with base tags
autoTaggingService.autoTagImage(image, bitmap, detectedFaces)
} catch (e: Exception) {
DiagnosticLogger.e("Base tag scan failed for $imageUri", e)
0
}
}
private suspend fun scanImageForSceneTags(
imageUri: String,
image: com.placeholder.sherpai2.data.local.entity.ImageEntity
): Int = withContext(Dispatchers.Default) {
try {
val uri = Uri.parse(imageUri)
val inputStream = getApplication<Application>().contentResolver.openInputStream(uri)
val bitmap = BitmapFactory.decodeStream(inputStream)
inputStream?.close()
if (bitmap == null) return@withContext 0
// Only auto-tag scene tags (indoor/outdoor already included in autoTagImage)
// This is a subset of base tags, so we don't need separate logic
0
} catch (e: Exception) {
DiagnosticLogger.e("Scene tag scan failed for $imageUri", e)
0
}
}
private suspend fun detectFaces(bitmap: android.graphics.Bitmap): List<DetectedFace> = withContext(Dispatchers.Default) {
try {
val image = InputImage.fromBitmap(bitmap, 0)
val faces = faceDetector.process(image).await()
faces.mapNotNull { face ->
val boundingBox = face.boundingBox
val croppedFace = try {
val left = boundingBox.left.coerceAtLeast(0)
val top = boundingBox.top.coerceAtLeast(0)
val width = boundingBox.width().coerceAtMost(bitmap.width - left)
val height = boundingBox.height().coerceAtMost(bitmap.height - top)
if (width > 0 && height > 0) {
android.graphics.Bitmap.createBitmap(bitmap, left, top, width, height)
} else {
null
}
} catch (e: Exception) {
null
}
if (croppedFace != null) {
DetectedFace(
croppedBitmap = croppedFace,
boundingBox = boundingBox
)
} else {
null
}
}
} catch (e: Exception) {
emptyList()
}
}
fun resetScanningState() {
_scanningState.value = TagScanningState.Idle
}
override fun onCleared() {
super.onCleared()
faceDetector.close()
}
}

View File

@@ -0,0 +1,77 @@
// TourScreen.kt
package com.placeholder.sherpai2.ui.tour
import androidx.compose.foundation.background
import androidx.compose.foundation.layout.*
import androidx.compose.foundation.lazy.LazyColumn
import androidx.compose.foundation.lazy.items
import androidx.compose.foundation.rememberScrollState
import androidx.compose.foundation.verticalScroll
import androidx.compose.material3.*
import androidx.compose.runtime.*
import androidx.compose.ui.Modifier
import androidx.compose.ui.unit.dp
import androidx.hilt.navigation.compose.hiltViewModel
import com.placeholder.sherpai2.data.local.model.ImageWithEverything
@Composable
fun TourScreen(tourViewModel: TourViewModel = hiltViewModel(), onImageClick: (String) -> Unit) {
val images by tourViewModel.recentImages.collectAsState()
Column(modifier = Modifier.fillMaxSize()) {
// Header with image count
Text(
text = "Gallery (${images.size} images)",
style = MaterialTheme.typography.titleLarge,
modifier = Modifier.padding(16.dp)
)
LazyColumn(
modifier = Modifier.fillMaxSize(),
contentPadding = PaddingValues(16.dp)
) {
items(images) { image ->
ImageCard(image)
Spacer(modifier = Modifier.height(12.dp))
}
}
}
}
@Composable
fun ImageCard(image: ImageWithEverything) {
Card(modifier = Modifier.fillMaxWidth(), elevation = CardDefaults.cardElevation(4.dp)) {
Column(modifier = Modifier.padding(12.dp)) {
Text(text = image.tags.toString(), style = MaterialTheme.typography.bodyMedium)
// Tags row with placeholders if fewer than 3
Row(modifier = Modifier.padding(top = 8.dp)) {
val tags = image.tags.map { it.tagId } // adjust depending on your entity
tags.forEach { tag ->
TagComposable(tag)
}
repeat(3 - tags.size.coerceAtMost(3)) {
TagComposable("") // empty placeholder
}
}
}
}
}
@Composable
fun TagComposable(tag: String) {
Box(
modifier = Modifier
.padding(end = 4.dp)
.height(24.dp)
.widthIn(min = 40.dp)
.background(MaterialTheme.colorScheme.primaryContainer, MaterialTheme.shapes.small),
contentAlignment = androidx.compose.ui.Alignment.Center
) {
Text(
text = if (tag.isNotBlank()) tag else " ",
style = MaterialTheme.typography.labelSmall,
modifier = Modifier.padding(horizontal = 6.dp)
)
}
}

View File

@@ -0,0 +1,39 @@
// TourViewModel.kt
package com.placeholder.sherpai2.ui.tour
import androidx.lifecycle.ViewModel
import androidx.lifecycle.viewModelScope
import com.placeholder.sherpai2.domain.repository.ImageRepository
import com.placeholder.sherpai2.data.local.model.ImageWithEverything
import dagger.hilt.android.lifecycle.HiltViewModel
import kotlinx.coroutines.flow.*
import kotlinx.coroutines.launch
import javax.inject.Inject
@HiltViewModel
class TourViewModel @Inject constructor(
private val imageRepository: ImageRepository
) : ViewModel() {
// Expose recent images as StateFlow
private val _recentImages = MutableStateFlow<List<ImageWithEverything>>(emptyList())
val recentImages: StateFlow<List<ImageWithEverything>> = _recentImages.asStateFlow()
init {
loadRecentImages()
}
private fun loadRecentImages(limit: Int = 100) {
viewModelScope.launch {
imageRepository.getRecentImages(limit)
.catch { e ->
println("TourViewModel: error fetching images: $e")
_recentImages.value = emptyList()
}
.collect { images ->
println("TourViewModel: fetched ${images.size} images")
_recentImages.value = images
}
}
}
}

View File

@@ -1,73 +0,0 @@
import androidx.compose.foundation.layout.Arrangement
import androidx.compose.foundation.layout.Box
import androidx.compose.foundation.layout.Column
import androidx.compose.foundation.layout.aspectRatio
import androidx.compose.foundation.layout.fillMaxSize
import androidx.compose.foundation.layout.fillMaxWidth
import androidx.compose.foundation.layout.padding
import androidx.compose.foundation.lazy.grid.GridCells
import androidx.compose.foundation.lazy.grid.LazyVerticalGrid
import androidx.compose.foundation.lazy.grid.items
import androidx.compose.material3.*
import androidx.compose.runtime.Composable
import androidx.compose.ui.Alignment
import androidx.compose.ui.Modifier
import androidx.compose.ui.layout.ContentScale
import androidx.compose.ui.unit.dp
import coil.compose.AsyncImage
import com.placeholder.sherpai2.data.photos.Photo
import com.placeholder.sherpai2.ui.tourscreen.GalleryUiState
@Composable
fun GalleryScreen(
state: GalleryUiState,
modifier: Modifier = Modifier // Add default modifier
) {
// Note: If this is inside MainContentArea, you might not need a second Scaffold.
// Let's use a Column or Box to ensure it fills the space correctly.
Column(
modifier = modifier.fillMaxSize()
) {
Text(
text = "Photo Gallery",
style = MaterialTheme.typography.headlineMedium,
modifier = Modifier.padding(16.dp)
)
Box(modifier = Modifier.fillMaxSize(), contentAlignment = Alignment.Center) {
when (state) {
is GalleryUiState.Loading -> CircularProgressIndicator()
is GalleryUiState.Error -> Text(text = state.message)
is GalleryUiState.Success -> {
if (state.photos.isEmpty()) {
Text("No photos found. Try adding some to the emulator!")
} else {
LazyVerticalGrid(
columns = GridCells.Fixed(3),
modifier = Modifier.fillMaxSize(),
horizontalArrangement = Arrangement.spacedBy(2.dp),
verticalArrangement = Arrangement.spacedBy(2.dp)
) {
items(state.photos) { photo ->
PhotoItem(photo)
}
}
}
}
}
}
}
}
@Composable
fun PhotoItem(photo: Photo) {
AsyncImage(
model = photo.uri,
contentDescription = photo.title,
modifier = Modifier
.aspectRatio(1f) // Makes it a square
.fillMaxWidth(),
contentScale = ContentScale.Crop
)
}

View File

@@ -1,9 +0,0 @@
package com.placeholder.sherpai2.ui.tourscreen
import com.placeholder.sherpai2.data.photos.Photo
sealed class GalleryUiState {
object Loading : GalleryUiState()
data class Success(val photos: List<Photo>) : GalleryUiState()
data class Error(val message: String) : GalleryUiState()
}

View File

@@ -1,35 +0,0 @@
import android.app.Application
import androidx.lifecycle.AndroidViewModel
import androidx.lifecycle.ViewModel
import androidx.lifecycle.viewModelScope
import com.placeholder.sherpai2.data.photos.Photo
import com.placeholder.sherpai2.data.repo.PhotoRepository
import com.placeholder.sherpai2.ui.tourscreen.GalleryUiState
import kotlinx.coroutines.flow.MutableStateFlow
import kotlinx.coroutines.flow.asStateFlow
import kotlinx.coroutines.launch
class GalleryViewModel(application: Application) : AndroidViewModel(application) {
// Initialize repository with the application context
private val repository = PhotoRepository(application)
private val _uiState = MutableStateFlow<GalleryUiState>(GalleryUiState.Loading)
val uiState = _uiState.asStateFlow()
init {
loadPhotos()
}
fun loadPhotos() {
viewModelScope.launch {
val result = repository.scanExternalStorage()
result.onSuccess { photos ->
_uiState.value = GalleryUiState.Success(photos)
}.onFailure { error ->
_uiState.value = GalleryUiState.Error(error.message ?: "Unknown Error")
}
}
}
}

View File

@@ -1,4 +0,0 @@
package com.placeholder.sherpai2.ui.tourscreen.components
class AlbumBox {
}

View File

@@ -0,0 +1,197 @@
package com.placeholder.sherpai2.ui.trainingprep
import android.os.Build
import android.view.View
import android.view.autofill.AutofillManager
import androidx.annotation.RequiresApi
import androidx.compose.foundation.layout.*
import androidx.compose.foundation.rememberScrollState
import androidx.compose.foundation.shape.RoundedCornerShape
import androidx.compose.foundation.verticalScroll
import androidx.compose.material.icons.Icons
import androidx.compose.material.icons.filled.*
import androidx.compose.material3.*
import androidx.compose.runtime.*
import androidx.compose.ui.Alignment
import androidx.compose.ui.Modifier
import androidx.compose.ui.platform.LocalView
import androidx.compose.ui.text.font.FontWeight
import androidx.compose.ui.text.input.KeyboardCapitalization
import androidx.compose.ui.unit.dp
import androidx.compose.ui.window.Dialog
import androidx.compose.ui.window.DialogProperties
import java.text.SimpleDateFormat
import java.util.*
@RequiresApi(Build.VERSION_CODES.O)
@OptIn(ExperimentalMaterial3Api::class)
@Composable
fun BeautifulPersonInfoDialog(
onDismiss: () -> Unit,
onConfirm: (name: String, dateOfBirth: Long?, relationship: String) -> Unit
) {
var name by remember { mutableStateOf("") }
var dateOfBirth by remember { mutableStateOf<Long?>(null) }
var selectedRelationship by remember { mutableStateOf("Other") }
var showDatePicker by remember { mutableStateOf(false) }
// ✅ Disable autofill for this dialog
val view = LocalView.current
DisposableEffect(Unit) {
val autofillManager = view.context.getSystemService(AutofillManager::class.java)
view.importantForAutofill = View.IMPORTANT_FOR_AUTOFILL_NO_EXCLUDE_DESCENDANTS
onDispose {
view.importantForAutofill = View.IMPORTANT_FOR_AUTOFILL_AUTO
}
}
val relationships = listOf(
"Family" to "👨‍👩‍👧‍👦",
"Friend" to "🤝",
"Partner" to "❤️",
"Parent" to "👪",
"Sibling" to "👫",
"Colleague" to "💼"
)
Dialog(
onDismissRequest = onDismiss,
properties = DialogProperties(usePlatformDefaultWidth = false)
) {
Card(
modifier = Modifier.fillMaxWidth(0.92f).fillMaxHeight(0.85f),
shape = RoundedCornerShape(28.dp),
colors = CardDefaults.cardColors(containerColor = MaterialTheme.colorScheme.surface),
elevation = CardDefaults.cardElevation(defaultElevation = 8.dp)
) {
Column(modifier = Modifier.fillMaxSize()) {
Row(
modifier = Modifier.fillMaxWidth().padding(24.dp),
horizontalArrangement = Arrangement.SpaceBetween,
verticalAlignment = Alignment.CenterVertically
) {
Row(horizontalArrangement = Arrangement.spacedBy(16.dp), verticalAlignment = Alignment.CenterVertically) {
Surface(shape = RoundedCornerShape(16.dp), color = MaterialTheme.colorScheme.primaryContainer, modifier = Modifier.size(64.dp)) {
Box(contentAlignment = Alignment.Center) {
Icon(Icons.Default.Person, contentDescription = null, modifier = Modifier.size(36.dp), tint = MaterialTheme.colorScheme.primary)
}
}
Column {
Text("Person Details", style = MaterialTheme.typography.headlineMedium, fontWeight = FontWeight.Bold)
Text("Help us organize your photos", style = MaterialTheme.typography.bodyMedium, color = MaterialTheme.colorScheme.onSurfaceVariant)
}
}
IconButton(onClick = onDismiss) {
Icon(Icons.Default.Close, contentDescription = "Close", modifier = Modifier.size(24.dp))
}
}
HorizontalDivider(color = MaterialTheme.colorScheme.outlineVariant)
Column(modifier = Modifier.weight(1f).verticalScroll(rememberScrollState()).padding(24.dp), verticalArrangement = Arrangement.spacedBy(24.dp)) {
Column(verticalArrangement = Arrangement.spacedBy(8.dp)) {
Text("Name *", style = MaterialTheme.typography.titleSmall, fontWeight = FontWeight.SemiBold, color = MaterialTheme.colorScheme.primary)
OutlinedTextField(
value = name,
onValueChange = { name = it },
placeholder = { Text("e.g., John Doe") },
leadingIcon = { Icon(Icons.Default.Face, contentDescription = null) },
modifier = Modifier.fillMaxWidth(),
singleLine = true,
shape = RoundedCornerShape(16.dp),
keyboardOptions = androidx.compose.foundation.text.KeyboardOptions(
capitalization = KeyboardCapitalization.Words,
autoCorrect = false
)
)
}
Column(verticalArrangement = Arrangement.spacedBy(8.dp)) {
Text("Birthday", style = MaterialTheme.typography.titleSmall, fontWeight = FontWeight.SemiBold, color = MaterialTheme.colorScheme.primary)
OutlinedTextField(
value = dateOfBirth?.let { SimpleDateFormat("MMM d, yyyy", Locale.getDefault()).format(Date(it)) } ?: "",
onValueChange = {},
readOnly = true,
placeholder = { Text("Select birthday") },
leadingIcon = { Icon(Icons.Default.Cake, contentDescription = null) },
trailingIcon = {
IconButton(onClick = { showDatePicker = true }) {
Icon(Icons.Default.CalendarToday, contentDescription = "Select date")
}
},
modifier = Modifier.fillMaxWidth(),
singleLine = true,
shape = RoundedCornerShape(16.dp)
)
}
Column(verticalArrangement = Arrangement.spacedBy(8.dp)) {
Text("Relationship", style = MaterialTheme.typography.titleSmall, fontWeight = FontWeight.SemiBold, color = MaterialTheme.colorScheme.primary)
var expanded by remember { mutableStateOf(false) }
ExposedDropdownMenuBox(expanded = expanded, onExpandedChange = { expanded = it }) {
OutlinedTextField(
value = selectedRelationship,
onValueChange = {},
readOnly = true,
leadingIcon = { Icon(Icons.Default.People, contentDescription = null) },
trailingIcon = { ExposedDropdownMenuDefaults.TrailingIcon(expanded = expanded) },
modifier = Modifier.fillMaxWidth().menuAnchor(),
singleLine = true,
shape = RoundedCornerShape(16.dp),
colors = ExposedDropdownMenuDefaults.outlinedTextFieldColors()
)
ExposedDropdownMenu(expanded = expanded, onDismissRequest = { expanded = false }) {
relationships.forEach { (relationship, emoji) ->
DropdownMenuItem(text = { Text("$emoji $relationship") }, onClick = { selectedRelationship = relationship; expanded = false })
}
}
}
}
Card(colors = CardDefaults.cardColors(containerColor = MaterialTheme.colorScheme.tertiaryContainer.copy(alpha = 0.3f)), shape = RoundedCornerShape(12.dp)) {
Row(modifier = Modifier.padding(16.dp), horizontalArrangement = Arrangement.spacedBy(12.dp)) {
Icon(Icons.Default.Lock, contentDescription = null, tint = MaterialTheme.colorScheme.tertiary, modifier = Modifier.size(20.dp))
Text("All information stays private on your device", style = MaterialTheme.typography.bodySmall, color = MaterialTheme.colorScheme.onTertiaryContainer)
}
}
}
HorizontalDivider(color = MaterialTheme.colorScheme.outlineVariant)
Row(modifier = Modifier.fillMaxWidth().padding(24.dp), horizontalArrangement = Arrangement.spacedBy(12.dp)) {
OutlinedButton(onClick = onDismiss, modifier = Modifier.weight(1f).height(56.dp), shape = RoundedCornerShape(16.dp)) {
Text("Cancel", style = MaterialTheme.typography.titleMedium)
}
Button(
onClick = { onConfirm(name.trim(), dateOfBirth, selectedRelationship) },
enabled = name.trim().isNotEmpty(),
modifier = Modifier.weight(1f).height(56.dp),
shape = RoundedCornerShape(16.dp)
) {
Icon(Icons.Default.Check, contentDescription = null, modifier = Modifier.size(20.dp))
Spacer(Modifier.width(8.dp))
Text("Continue", style = MaterialTheme.typography.titleMedium, fontWeight = FontWeight.Bold)
}
}
}
}
}
if (showDatePicker) {
val datePickerState = rememberDatePickerState(initialSelectedDateMillis = dateOfBirth ?: System.currentTimeMillis())
DatePickerDialog(
onDismissRequest = { showDatePicker = false },
confirmButton = { TextButton(onClick = { dateOfBirth = datePickerState.selectedDateMillis; showDatePicker = false }) { Text("OK") } },
dismissButton = { TextButton(onClick = { showDatePicker = false }) { Text("Cancel") } }
) {
DatePicker(state = datePickerState)
}
}
}

View File

@@ -0,0 +1,159 @@
package com.placeholder.sherpai2.ui.trainingprep
import android.content.Context
import android.graphics.Bitmap
import android.graphics.BitmapFactory
import android.net.Uri
import kotlinx.coroutines.Dispatchers
import kotlinx.coroutines.withContext
import java.io.InputStream
/**
* Helper class for detecting duplicate or near-duplicate images using perceptual hashing
*/
class DuplicateImageDetector(private val context: Context) {
data class DuplicateCheckResult(
val hasDuplicates: Boolean,
val duplicateGroups: List<DuplicateGroup>,
val uniqueImageCount: Int
)
data class DuplicateGroup(
val images: List<Uri>,
val similarity: Double
)
private data class ImageHash(
val uri: Uri,
val hash: Long
)
/**
* Check for duplicate images in the provided list
*/
suspend fun checkForDuplicates(
uris: List<Uri>,
similarityThreshold: Double = 0.95
): DuplicateCheckResult = withContext(Dispatchers.Default) {
if (uris.size < 2) {
return@withContext DuplicateCheckResult(
hasDuplicates = false,
duplicateGroups = emptyList(),
uniqueImageCount = uris.size
)
}
// Compute perceptual hash for each image
val imageHashes = uris.mapNotNull { uri ->
try {
val bitmap = loadBitmap(uri)
bitmap?.let {
val hash = computePerceptualHash(it)
ImageHash(uri, hash)
}
} catch (e: Exception) {
null
}
}
// Find duplicate groups
val duplicateGroups = mutableListOf<DuplicateGroup>()
val processed = mutableSetOf<Uri>()
for (i in imageHashes.indices) {
if (imageHashes[i].uri in processed) continue
val currentGroup = mutableListOf(imageHashes[i].uri)
for (j in i + 1 until imageHashes.size) {
if (imageHashes[j].uri in processed) continue
val similarity = calculateSimilarity(imageHashes[i].hash, imageHashes[j].hash)
if (similarity >= similarityThreshold) {
currentGroup.add(imageHashes[j].uri)
processed.add(imageHashes[j].uri)
}
}
if (currentGroup.size > 1) {
duplicateGroups.add(
DuplicateGroup(
images = currentGroup,
similarity = 1.0
)
)
processed.addAll(currentGroup)
}
}
DuplicateCheckResult(
hasDuplicates = duplicateGroups.isNotEmpty(),
duplicateGroups = duplicateGroups,
uniqueImageCount = uris.size - duplicateGroups.sumOf { it.images.size - 1 }
)
}
/**
* Compute perceptual hash using difference hash (dHash) algorithm
*/
private fun computePerceptualHash(bitmap: Bitmap): Long {
// Resize to 9x8
val resized = Bitmap.createScaledBitmap(bitmap, 9, 8, false)
var hash = 0L
var bitIndex = 0
for (y in 0 until 8) {
for (x in 0 until 8) {
val leftPixel = resized.getPixel(x, y)
val rightPixel = resized.getPixel(x + 1, y)
val leftGray = toGrayscale(leftPixel)
val rightGray = toGrayscale(rightPixel)
if (leftGray > rightGray) {
hash = hash or (1L shl bitIndex)
}
bitIndex++
}
}
resized.recycle()
return hash
}
/**
* Convert RGB pixel to grayscale value
*/
private fun toGrayscale(pixel: Int): Int {
val r = (pixel shr 16) and 0xFF
val g = (pixel shr 8) and 0xFF
val b = pixel and 0xFF
return (0.299 * r + 0.587 * g + 0.114 * b).toInt()
}
/**
* Calculate similarity between two hashes
*/
private fun calculateSimilarity(hash1: Long, hash2: Long): Double {
val xor = hash1 xor hash2
val hammingDistance = xor.countOneBits()
return 1.0 - (hammingDistance / 64.0)
}
/**
* Load bitmap from URI
*/
private fun loadBitmap(uri: Uri): Bitmap? {
return try {
val inputStream: InputStream? = context.contentResolver.openInputStream(uri)
BitmapFactory.decodeStream(inputStream)?.also {
inputStream?.close()
}
} catch (e: Exception) {
null
}
}
}

View File

@@ -0,0 +1,360 @@
package com.placeholder.sherpai2.ui.trainingprep
import android.net.Uri
import androidx.compose.foundation.BorderStroke
import androidx.compose.foundation.background
import androidx.compose.foundation.layout.*
import androidx.compose.foundation.lazy.LazyRow
import androidx.compose.foundation.lazy.items
import androidx.compose.foundation.shape.RoundedCornerShape
import androidx.compose.material.icons.Icons
import androidx.compose.material.icons.filled.*
import androidx.compose.material3.*
import androidx.compose.runtime.*
import androidx.compose.ui.Alignment
import androidx.compose.ui.Modifier
import androidx.compose.ui.graphics.Color
import androidx.compose.ui.layout.ContentScale
import androidx.compose.ui.text.font.FontWeight
import androidx.compose.ui.unit.dp
import coil.compose.AsyncImage
/**
* DuplicateImageHighlighter - Enhanced duplicate detection UI
*
* FEATURES:
* - Visual highlighting of duplicate groups
* - Shows thumbnail previews of duplicates
* - One-click "Remove Duplicate" button
* - Keeps best image automatically
* - Warning badge with count
*
* GENTLE UX:
* - Non-intrusive warning color (amber, not red)
* - Clear visual grouping
* - Simple action ("Remove" vs "Keep")
* - Automatic selection of which to remove
*/
@Composable
fun DuplicateImageHighlighter(
duplicateGroups: List<DuplicateImageDetector.DuplicateGroup>,
allImageUris: List<Uri>,
onRemoveDuplicate: (Uri) -> Unit,
modifier: Modifier = Modifier
) {
if (duplicateGroups.isEmpty()) return
Column(
modifier = modifier
.fillMaxWidth()
.padding(vertical = 8.dp),
verticalArrangement = Arrangement.spacedBy(12.dp)
) {
// Header with count
Row(
modifier = Modifier.fillMaxWidth(),
horizontalArrangement = Arrangement.SpaceBetween,
verticalAlignment = Alignment.CenterVertically
) {
Row(
horizontalArrangement = Arrangement.spacedBy(8.dp),
verticalAlignment = Alignment.CenterVertically
) {
Icon(
Icons.Default.Warning,
contentDescription = null,
tint = MaterialTheme.colorScheme.tertiary, // Amber, not red
modifier = Modifier.size(20.dp)
)
Text(
"${duplicateGroups.size} duplicate ${if (duplicateGroups.size == 1) "group" else "groups"} found",
style = MaterialTheme.typography.titleMedium,
fontWeight = FontWeight.Bold
)
}
// Total duplicates badge
Surface(
shape = RoundedCornerShape(12.dp),
color = MaterialTheme.colorScheme.tertiaryContainer
) {
Text(
"${duplicateGroups.sumOf { it.images.size - 1 }} to remove",
modifier = Modifier.padding(horizontal = 12.dp, vertical = 4.dp),
style = MaterialTheme.typography.labelMedium,
color = MaterialTheme.colorScheme.onTertiaryContainer,
fontWeight = FontWeight.Bold
)
}
}
// Each duplicate group
duplicateGroups.forEachIndexed { groupIndex, group ->
DuplicateGroupCard(
groupIndex = groupIndex + 1,
duplicateGroup = group,
onRemove = onRemoveDuplicate
)
}
}
}
/**
* Card showing one duplicate group with thumbnails
*/
@Composable
private fun DuplicateGroupCard(
groupIndex: Int,
duplicateGroup: DuplicateImageDetector.DuplicateGroup,
onRemove: (Uri) -> Unit
) {
var expanded by remember { mutableStateOf(false) }
Card(
modifier = Modifier.fillMaxWidth(),
colors = CardDefaults.cardColors(
containerColor = MaterialTheme.colorScheme.tertiaryContainer.copy(alpha = 0.3f)
),
border = BorderStroke(1.dp, MaterialTheme.colorScheme.tertiary.copy(alpha = 0.3f)),
shape = RoundedCornerShape(12.dp)
) {
Column(
modifier = Modifier
.fillMaxWidth()
.padding(12.dp),
verticalArrangement = Arrangement.spacedBy(12.dp)
) {
// Header row
Row(
modifier = Modifier.fillMaxWidth(),
horizontalArrangement = Arrangement.SpaceBetween,
verticalAlignment = Alignment.CenterVertically
) {
Row(
horizontalArrangement = Arrangement.spacedBy(8.dp),
verticalAlignment = Alignment.CenterVertically
) {
// Group number badge
Surface(
shape = RoundedCornerShape(8.dp),
color = MaterialTheme.colorScheme.tertiary
) {
Text(
"#$groupIndex",
modifier = Modifier.padding(horizontal = 8.dp, vertical = 4.dp),
style = MaterialTheme.typography.labelMedium,
color = MaterialTheme.colorScheme.onTertiary,
fontWeight = FontWeight.Bold
)
}
Text(
"${duplicateGroup.images.size} identical images",
style = MaterialTheme.typography.bodyMedium,
fontWeight = FontWeight.SemiBold
)
}
// Expand/collapse button
IconButton(
onClick = { expanded = !expanded },
modifier = Modifier.size(32.dp)
) {
Icon(
if (expanded) Icons.Default.ExpandLess else Icons.Default.ExpandMore,
contentDescription = if (expanded) "Collapse" else "Expand"
)
}
}
// Thumbnail row (always visible)
LazyRow(
horizontalArrangement = Arrangement.spacedBy(8.dp)
) {
items(duplicateGroup.images.take(3)) { uri ->
DuplicateThumbnail(
uri = uri,
similarity = duplicateGroup.similarity
)
}
if (duplicateGroup.images.size > 3) {
item {
Surface(
modifier = Modifier
.size(80.dp),
shape = RoundedCornerShape(8.dp),
color = MaterialTheme.colorScheme.surfaceVariant
) {
Box(contentAlignment = Alignment.Center) {
Text(
"+${duplicateGroup.images.size - 3}",
style = MaterialTheme.typography.titleMedium,
fontWeight = FontWeight.Bold
)
}
}
}
}
}
// Action buttons
Row(
modifier = Modifier.fillMaxWidth(),
horizontalArrangement = Arrangement.spacedBy(8.dp)
) {
// Keep first, remove rest
Button(
onClick = {
// Remove all but the first image
duplicateGroup.images.drop(1).forEach { uri ->
onRemove(uri)
}
},
modifier = Modifier.weight(1f),
colors = ButtonDefaults.buttonColors(
containerColor = MaterialTheme.colorScheme.tertiary
)
) {
Icon(
Icons.Default.DeleteSweep,
contentDescription = null,
modifier = Modifier.size(18.dp)
)
Spacer(Modifier.width(6.dp))
Text("Remove ${duplicateGroup.images.size - 1} Duplicates")
}
}
// Expanded info (optional)
if (expanded) {
HorizontalDivider(color = MaterialTheme.colorScheme.outline.copy(alpha = 0.3f))
Column(
verticalArrangement = Arrangement.spacedBy(8.dp)
) {
Text(
"Individual actions:",
style = MaterialTheme.typography.labelMedium,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
duplicateGroup.images.forEachIndexed { index, uri ->
Row(
modifier = Modifier.fillMaxWidth(),
horizontalArrangement = Arrangement.SpaceBetween,
verticalAlignment = Alignment.CenterVertically
) {
Row(
horizontalArrangement = Arrangement.spacedBy(8.dp),
verticalAlignment = Alignment.CenterVertically,
modifier = Modifier.weight(1f)
) {
AsyncImage(
model = uri,
contentDescription = null,
modifier = Modifier
.size(40.dp)
.background(
MaterialTheme.colorScheme.surfaceVariant,
RoundedCornerShape(6.dp)
),
contentScale = ContentScale.Crop
)
Text(
uri.lastPathSegment?.take(20) ?: "Image ${index + 1}",
style = MaterialTheme.typography.bodySmall,
modifier = Modifier.weight(1f)
)
}
if (index == 0) {
// First image - will be kept
Surface(
shape = RoundedCornerShape(8.dp),
color = MaterialTheme.colorScheme.primaryContainer
) {
Row(
modifier = Modifier.padding(horizontal = 8.dp, vertical = 4.dp),
horizontalArrangement = Arrangement.spacedBy(4.dp),
verticalAlignment = Alignment.CenterVertically
) {
Icon(
Icons.Default.CheckCircle,
contentDescription = null,
modifier = Modifier.size(14.dp),
tint = MaterialTheme.colorScheme.primary
)
Text(
"Keep",
style = MaterialTheme.typography.labelSmall,
color = MaterialTheme.colorScheme.primary,
fontWeight = FontWeight.Bold
)
}
}
} else {
// Duplicate - will be removed
TextButton(
onClick = { onRemove(uri) },
colors = ButtonDefaults.textButtonColors(
contentColor = MaterialTheme.colorScheme.error
)
) {
Icon(
Icons.Default.Delete,
contentDescription = null,
modifier = Modifier.size(16.dp)
)
Spacer(Modifier.width(4.dp))
Text("Remove", style = MaterialTheme.typography.labelMedium)
}
}
}
}
}
}
}
}
}
/**
* Thumbnail with similarity badge
*/
@Composable
private fun DuplicateThumbnail(
uri: Uri,
similarity: Double
) {
Box {
AsyncImage(
model = uri,
contentDescription = null,
modifier = Modifier
.size(80.dp)
.background(
MaterialTheme.colorScheme.surfaceVariant,
RoundedCornerShape(8.dp)
),
contentScale = ContentScale.Crop
)
// Similarity badge
Surface(
modifier = Modifier
.align(Alignment.BottomEnd)
.padding(4.dp),
shape = RoundedCornerShape(4.dp),
color = MaterialTheme.colorScheme.tertiaryContainer.copy(alpha = 0.9f)
) {
Text(
"${(similarity * 100).toInt()}%",
modifier = Modifier.padding(horizontal = 4.dp, vertical = 2.dp),
style = MaterialTheme.typography.labelSmall,
color = MaterialTheme.colorScheme.onTertiaryContainer,
fontWeight = FontWeight.Bold
)
}
}
}

View File

@@ -0,0 +1,340 @@
package com.placeholder.sherpai2.ui.trainingprep
import android.graphics.Bitmap
import android.graphics.BitmapFactory
import android.graphics.Rect
import android.net.Uri
import androidx.compose.foundation.BorderStroke
import androidx.compose.foundation.Image
import androidx.compose.foundation.clickable
import androidx.compose.foundation.layout.*
import androidx.compose.foundation.shape.CircleShape
import androidx.compose.foundation.shape.RoundedCornerShape
import androidx.compose.material.icons.Icons
import androidx.compose.material.icons.filled.*
import androidx.compose.material3.*
import androidx.compose.runtime.*
import androidx.compose.ui.Alignment
import androidx.compose.ui.Modifier
import androidx.compose.ui.graphics.asImageBitmap
import androidx.compose.ui.layout.ContentScale
import androidx.compose.ui.platform.LocalContext
import androidx.compose.ui.text.font.FontWeight
import androidx.compose.ui.unit.dp
import androidx.compose.ui.window.Dialog
import androidx.compose.ui.window.DialogProperties
import com.google.mlkit.vision.common.InputImage
import com.google.mlkit.vision.face.FaceDetection
import androidx.compose.ui.graphics.Color
import com.google.mlkit.vision.face.FaceDetectorOptions
import kotlinx.coroutines.Dispatchers
import kotlinx.coroutines.tasks.await
import kotlinx.coroutines.withContext
/**
* MINIMAL FacePickerDialog - Optimized for batch processing 30-50 photos
*
* REMOVED CLUTTER:
* - "Preview (tap to select)" header
* - "Face will be used for training" info box
* - "Face #" labels covering previews
* - Original image preview
*
* IMPROVED:
* - Larger face previews (1:1 aspect ratio)
* - Clean checkmark overlay only
* - Minimal text
* - Fast workflow
*/
@Composable
fun FacePickerDialog(
result: FaceDetectionHelper.FaceDetectionResult,
onDismiss: () -> Unit,
onFaceSelected: (Int, Bitmap) -> Unit
) {
val context = LocalContext.current
var selectedFaceIndex by remember { mutableStateOf(0) }
var croppedFaces by remember { mutableStateOf<List<Bitmap>>(emptyList()) }
var isLoading by remember { mutableStateOf(true) }
var errorMessage by remember { mutableStateOf<String?>(null) }
// Load and crop all faces - RE-DETECT to get accurate bounds
LaunchedEffect(result) {
isLoading = true
errorMessage = null
try {
croppedFaces = withContext(Dispatchers.IO) {
// Load the FULL resolution bitmap (no downsampling)
val fullBitmap = loadFullResolutionBitmap(context, result.uri)
if (fullBitmap == null) {
errorMessage = "Failed to load image"
return@withContext emptyList()
}
// Re-detect faces on the full resolution bitmap to get accurate bounds
val accurateFaceBounds = detectFacesOnBitmap(fullBitmap)
if (accurateFaceBounds.isEmpty()) {
// Fallback: try to use the original bounds with scaling
val scaledBounds = result.faceBounds.map { originalBounds ->
cropFaceFromBitmap(fullBitmap, originalBounds)
}
fullBitmap.recycle()
return@withContext scaledBounds
}
// Crop faces using accurate bounds
val croppedList = accurateFaceBounds.map { bounds ->
cropFaceFromBitmap(fullBitmap, bounds)
}
// CRITICAL: Recycle AFTER all cropping is done
fullBitmap.recycle()
croppedList
}
if (croppedFaces.isEmpty() && errorMessage == null) {
errorMessage = "No faces found in full resolution image"
}
} catch (e: Exception) {
errorMessage = "Error processing faces: ${e.message}"
} finally {
isLoading = false
}
}
Dialog(
onDismissRequest = onDismiss,
properties = DialogProperties(usePlatformDefaultWidth = false)
) {
Card(
modifier = Modifier
.fillMaxWidth(0.92f)
.wrapContentHeight(),
shape = RoundedCornerShape(20.dp),
colors = CardDefaults.cardColors(
containerColor = MaterialTheme.colorScheme.surface
)
) {
Column(
modifier = Modifier
.fillMaxWidth()
.padding(20.dp),
verticalArrangement = Arrangement.spacedBy(16.dp)
) {
// Minimal header - just close button
Row(
modifier = Modifier.fillMaxWidth(),
horizontalArrangement = Arrangement.SpaceBetween,
verticalAlignment = Alignment.CenterVertically
) {
Text(
text = "${result.faceCount} faces",
style = MaterialTheme.typography.titleLarge,
fontWeight = FontWeight.Bold
)
IconButton(onClick = onDismiss) {
Icon(Icons.Default.Close, contentDescription = "Close")
}
}
if (isLoading) {
// Loading state - minimal
Box(
modifier = Modifier
.fillMaxWidth()
.height(180.dp),
contentAlignment = Alignment.Center
) {
CircularProgressIndicator()
}
} else if (errorMessage != null) {
// Error state - minimal
Text(
text = errorMessage!!,
color = MaterialTheme.colorScheme.error,
style = MaterialTheme.typography.bodyMedium
)
} else {
// CLEAN face grid - NO labels, NO text
Row(
modifier = Modifier.fillMaxWidth(),
horizontalArrangement = Arrangement.spacedBy(10.dp)
) {
croppedFaces.forEachIndexed { index, faceBitmap ->
CleanFaceCard(
faceBitmap = faceBitmap,
isSelected = selectedFaceIndex == index,
onClick = { selectedFaceIndex = index },
modifier = Modifier.weight(1f)
)
}
}
}
// Action buttons - minimal
Row(
modifier = Modifier.fillMaxWidth(),
horizontalArrangement = Arrangement.spacedBy(10.dp)
) {
TextButton(
onClick = onDismiss,
modifier = Modifier.weight(1f)
) {
Text("Skip")
}
Button(
onClick = {
if (selectedFaceIndex < croppedFaces.size) {
onFaceSelected(selectedFaceIndex, croppedFaces[selectedFaceIndex])
}
},
enabled = !isLoading && croppedFaces.isNotEmpty(),
modifier = Modifier.weight(1f)
) {
Icon(
Icons.Default.Check,
contentDescription = null,
modifier = Modifier.size(18.dp)
)
Spacer(modifier = Modifier.width(6.dp))
Text("Use")
}
}
}
}
}
}
/**
* ULTRA-CLEAN face card - NO TEXT, just image + checkmark
*
* CHANGES:
* - 1:1 aspect ratio (bigger!)
* - NO "Face #" label
* - Checkmark in corner only
* - Minimal border
*/
@Composable
private fun CleanFaceCard(
faceBitmap: Bitmap,
isSelected: Boolean,
onClick: () -> Unit,
modifier: Modifier = Modifier
) {
Card(
modifier = modifier
.aspectRatio(1f) // SQUARE = bigger previews!
.clickable(onClick = onClick),
colors = CardDefaults.cardColors(
containerColor = MaterialTheme.colorScheme.surfaceVariant
),
border = if (isSelected)
BorderStroke(3.dp, MaterialTheme.colorScheme.primary)
else
BorderStroke(1.dp, MaterialTheme.colorScheme.outline.copy(alpha = 0.3f)),
shape = RoundedCornerShape(12.dp),
elevation = CardDefaults.cardElevation(
defaultElevation = if (isSelected) 4.dp else 1.dp
)
) {
Box(modifier = Modifier.fillMaxSize()) {
// Face image - FULL SIZE
Image(
bitmap = faceBitmap.asImageBitmap(),
contentDescription = null,
modifier = Modifier.fillMaxSize(),
contentScale = ContentScale.Crop
)
// Checkmark in corner - ONLY if selected
if (isSelected) {
Surface(
modifier = Modifier
.align(Alignment.TopEnd)
.padding(6.dp)
.size(32.dp),
shape = CircleShape,
color = MaterialTheme.colorScheme.primary,
shadowElevation = 4.dp
) {
Icon(
Icons.Default.CheckCircle,
contentDescription = "Selected",
modifier = Modifier
.padding(6.dp)
.size(20.dp),
tint = MaterialTheme.colorScheme.onPrimary
)
}
}
}
}
}
/**
* Load full resolution bitmap WITHOUT downsampling
*/
private suspend fun loadFullResolutionBitmap(
context: android.content.Context,
uri: Uri
): Bitmap? = withContext(Dispatchers.IO) {
try {
val inputStream = context.contentResolver.openInputStream(uri)
BitmapFactory.decodeStream(inputStream)?.also {
inputStream?.close()
}
} catch (e: Exception) {
null
}
}
/**
* Re-detect faces on full resolution bitmap to get accurate bounds
*/
private suspend fun detectFacesOnBitmap(bitmap: Bitmap): List<Rect> = withContext(Dispatchers.Default) {
try {
val options = FaceDetectorOptions.Builder()
.setPerformanceMode(FaceDetectorOptions.PERFORMANCE_MODE_ACCURATE)
.setLandmarkMode(FaceDetectorOptions.LANDMARK_MODE_NONE)
.setClassificationMode(FaceDetectorOptions.CLASSIFICATION_MODE_NONE)
.setMinFaceSize(0.10f)
.build()
val detector = FaceDetection.getClient(options)
val image = InputImage.fromBitmap(bitmap, 0)
val faces = detector.process(image).await()
// Sort by size (largest first)
faces.sortedByDescending { face ->
face.boundingBox.width() * face.boundingBox.height()
}.map { it.boundingBox }
} catch (e: Exception) {
emptyList()
}
}
/**
* Crop face from bitmap with padding
*/
private fun cropFaceFromBitmap(bitmap: Bitmap, faceBounds: Rect): Bitmap {
// Add 20% padding around the face
val padding = (faceBounds.width() * 0.2f).toInt()
val left = (faceBounds.left - padding).coerceAtLeast(0)
val top = (faceBounds.top - padding).coerceAtLeast(0)
val right = (faceBounds.right + padding).coerceAtMost(bitmap.width)
val bottom = (faceBounds.bottom + padding).coerceAtMost(bitmap.height)
val width = right - left
val height = bottom - top
return Bitmap.createBitmap(bitmap, left, top, width, height)
}

View File

@@ -0,0 +1,57 @@
package com.placeholder.sherpai2.ui.trainingprep
import android.graphics.Rect
import android.net.Uri
import androidx.compose.foundation.Canvas
import androidx.compose.foundation.Image
import androidx.compose.foundation.clickable
import androidx.compose.foundation.layout.*
import androidx.compose.foundation.lazy.LazyRow
import androidx.compose.foundation.lazy.items
import androidx.compose.material3.Card
import androidx.compose.material3.MaterialTheme
import androidx.compose.material3.Text
import androidx.compose.runtime.Composable
import androidx.compose.ui.Alignment
import androidx.compose.ui.Modifier
import androidx.compose.ui.layout.ContentScale
import androidx.compose.ui.unit.dp
import coil.compose.rememberAsyncImagePainter
@Composable
fun FacePickerScreen(
uri: Uri,
faceBoxes: List<Rect>,
onFaceSelected: (Rect) -> Unit
) {
Column(modifier = Modifier.fillMaxSize().padding(16.dp)) {
Text("Multiple faces detected!", style = MaterialTheme.typography.headlineSmall)
Text("Tap the person you want to train on.")
Box(modifier = Modifier.weight(1f).fillMaxWidth().padding(vertical = 16.dp)) {
// Main Image
Image(
painter = rememberAsyncImagePainter(uri),
contentDescription = null,
modifier = Modifier.fillMaxSize(),
contentScale = ContentScale.Fit
)
// Overlay Clickable Boxes
// Note: In a production app, you'd need to map Rect coordinates
// from the Bitmap scale to the UI View scale.
Canvas(modifier = Modifier.fillMaxSize().clickable { /* Handle general tap */ }) {
// Implementation of coordinate mapping goes here
// TODO implement coordinate mapping
}
// Simplified: Just show the options as a list of crops if Canvas mapping is too complex for now
LazyRow {
items(faceBoxes) { box ->
Card(modifier = Modifier.padding(8.dp).size(100.dp).clickable { onFaceSelected(box) }) {
Text("Face ${faceBoxes.indexOf(box) + 1}", Modifier.align(Alignment.CenterHorizontally))
}
}
}
}
}
}

View File

@@ -0,0 +1,185 @@
package com.placeholder.sherpai2.ui.trainingprep
import android.content.Context
import android.graphics.Bitmap
import android.graphics.BitmapFactory
import android.graphics.Rect
import android.net.Uri
import com.google.mlkit.vision.common.InputImage
import com.google.mlkit.vision.face.FaceDetection
import com.google.mlkit.vision.face.FaceDetectorOptions
import kotlinx.coroutines.Dispatchers
import kotlinx.coroutines.async
import kotlinx.coroutines.awaitAll
import kotlinx.coroutines.coroutineScope
import kotlinx.coroutines.tasks.await
import kotlinx.coroutines.withContext
import java.io.InputStream
/**
* FIXED FaceDetectionHelper with parallel processing
*
* FIXES:
* - Removed bitmap.recycle() that broke face cropping
* - Proper memory management with downsampling
* - Parallel processing for speed
*/
class FaceDetectionHelper(private val context: Context) {
private val faceDetectorOptions = FaceDetectorOptions.Builder()
.setPerformanceMode(FaceDetectorOptions.PERFORMANCE_MODE_ACCURATE) // ACCURATE for quality
.setLandmarkMode(FaceDetectorOptions.LANDMARK_MODE_ALL)
.setClassificationMode(FaceDetectorOptions.CLASSIFICATION_MODE_ALL)
.setMinFaceSize(0.15f)
.build()
private val detector = FaceDetection.getClient(faceDetectorOptions)
data class FaceDetectionResult(
val uri: Uri,
val hasFace: Boolean,
val faceCount: Int,
val faceBounds: List<Rect> = emptyList(),
val croppedFaceBitmap: Bitmap? = null, // Only largest face
val errorMessage: String? = null
)
/**
* Detect faces in a single image
*/
suspend fun detectFacesInImage(uri: Uri): FaceDetectionResult {
return withContext(Dispatchers.IO) {
var bitmap: Bitmap? = null
try {
bitmap = loadBitmap(uri)
if (bitmap == null) {
return@withContext FaceDetectionResult(
uri = uri,
hasFace = false,
faceCount = 0,
errorMessage = "Failed to load image"
)
}
val inputImage = InputImage.fromBitmap(bitmap, 0)
val faces = detector.process(inputImage).await()
// Sort by face size (area) to get the largest face
val sortedFaces = faces.sortedByDescending { face ->
face.boundingBox.width() * face.boundingBox.height()
}
val croppedFace = if (sortedFaces.isNotEmpty()) {
// Crop the LARGEST detected face (most likely the subject)
cropFaceFromBitmap(bitmap, sortedFaces[0].boundingBox)
} else null
FaceDetectionResult(
uri = uri,
hasFace = faces.isNotEmpty(),
faceCount = faces.size,
faceBounds = faces.map { it.boundingBox },
croppedFaceBitmap = croppedFace
)
} catch (e: Exception) {
FaceDetectionResult(
uri = uri,
hasFace = false,
faceCount = 0,
errorMessage = e.message ?: "Unknown error"
)
} finally {
// NOW we can recycle after we're completely done
bitmap?.recycle()
}
}
}
/**
* PARALLEL face detection in multiple images - 10x FASTER!
*
* @param onProgress Callback with (current, total)
*/
suspend fun detectFacesInImages(
uris: List<Uri>,
onProgress: ((Int, Int) -> Unit)? = null
): List<FaceDetectionResult> = coroutineScope {
val total = uris.size
var completed = 0
// Process in parallel batches of 5 to avoid overwhelming the system
uris.chunked(5).flatMap { batch ->
batch.map { uri ->
async(Dispatchers.IO) {
val result = detectFacesInImage(uri)
synchronized(this@FaceDetectionHelper) {
completed++
onProgress?.invoke(completed, total)
}
result
}
}.awaitAll()
}
}
/**
* Crop face from bitmap with padding
*/
private fun cropFaceFromBitmap(bitmap: Bitmap, faceBounds: Rect): Bitmap {
// Add 20% padding around the face
val padding = (faceBounds.width() * 0.2f).toInt()
val left = (faceBounds.left - padding).coerceAtLeast(0)
val top = (faceBounds.top - padding).coerceAtLeast(0)
val right = (faceBounds.right + padding).coerceAtMost(bitmap.width)
val bottom = (faceBounds.bottom + padding).coerceAtMost(bitmap.height)
val width = right - left
val height = bottom - top
return Bitmap.createBitmap(bitmap, left, top, width, height)
}
/**
* Load bitmap from URI with downsampling for memory efficiency
*/
private fun loadBitmap(uri: Uri): Bitmap? {
return try {
val inputStream: InputStream? = context.contentResolver.openInputStream(uri)
// First decode with inJustDecodeBounds to get dimensions
val options = BitmapFactory.Options().apply {
inJustDecodeBounds = true
}
BitmapFactory.decodeStream(inputStream, null, options)
inputStream?.close()
// Calculate sample size to limit max dimension to 1024px
val maxDimension = 1024
var sampleSize = 1
while (options.outWidth / sampleSize > maxDimension ||
options.outHeight / sampleSize > maxDimension) {
sampleSize *= 2
}
// Now decode with sample size
val inputStream2 = context.contentResolver.openInputStream(uri)
val finalOptions = BitmapFactory.Options().apply {
inSampleSize = sampleSize
}
BitmapFactory.decodeStream(inputStream2, null, finalOptions)?.also {
inputStream2?.close()
}
} catch (e: Exception) {
null
}
}
/**
* Clean up resources
*/
fun cleanup() {
detector.close()
}
}

View File

@@ -0,0 +1,403 @@
package com.placeholder.sherpai2.ui.trainingprep
import android.net.Uri
import androidx.activity.compose.rememberLauncherForActivityResult
import androidx.activity.result.contract.ActivityResultContracts
import androidx.compose.animation.AnimatedVisibility
import androidx.compose.foundation.layout.*
import androidx.compose.foundation.rememberScrollState
import androidx.compose.foundation.shape.RoundedCornerShape
import androidx.compose.foundation.verticalScroll
import androidx.compose.material.icons.Icons
import androidx.compose.material.icons.filled.*
import androidx.compose.material3.*
import androidx.compose.runtime.*
import androidx.compose.ui.Alignment
import androidx.compose.ui.Modifier
import androidx.compose.ui.graphics.Color
import androidx.compose.ui.text.font.FontWeight
import androidx.compose.ui.unit.dp
import androidx.hilt.navigation.compose.hiltViewModel
import androidx.lifecycle.compose.collectAsStateWithLifecycle
import com.placeholder.sherpai2.data.local.dao.ImageDao
import com.placeholder.sherpai2.data.local.entity.ImageEntity
import kotlinx.coroutines.launch
/**
* OPTIMIZED ImageSelectorScreen
*
* 🎯 NEW FEATURE: Filter to only show face-tagged images!
* ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━
* - Uses face detection cache to pre-filter
* - Shows "Only photos with faces" toggle
* - Dramatically faster photo selection
* - Better training quality (no manual filtering needed)
*/
@OptIn(ExperimentalMaterial3Api::class)
@Composable
fun ImageSelectorScreen(
onImagesSelected: (List<Uri>) -> Unit
) {
// Inject ImageDao via Hilt ViewModel pattern
val viewModel: ImageSelectorViewModel = hiltViewModel()
val faceTaggedUris by viewModel.faceTaggedImageUris.collectAsStateWithLifecycle()
var selectedImages by remember { mutableStateOf<List<Uri>>(emptyList()) }
var onlyShowFaceImages by remember { mutableStateOf(true) } // Default: smart filtering
val scrollState = rememberScrollState()
val photoPicker = rememberLauncherForActivityResult(
contract = ActivityResultContracts.GetMultipleContents()
) { uris ->
if (uris.isNotEmpty()) {
// Filter to only face-tagged images if toggle is on
selectedImages = if (onlyShowFaceImages && faceTaggedUris.isNotEmpty()) {
uris.filter { it.toString() in faceTaggedUris }
} else {
uris
}
}
}
Scaffold(
topBar = {
TopAppBar(
title = { Text("Select Training Photos") },
colors = TopAppBarDefaults.topAppBarColors(
containerColor = MaterialTheme.colorScheme.primaryContainer
)
)
}
) { paddingValues ->
Column(
modifier = Modifier
.fillMaxSize()
.padding(paddingValues)
.verticalScroll(scrollState)
.padding(16.dp),
verticalArrangement = Arrangement.spacedBy(16.dp)
) {
// Smart filtering card
if (faceTaggedUris.isNotEmpty()) {
Card(
modifier = Modifier.fillMaxWidth(),
colors = CardDefaults.cardColors(
containerColor = MaterialTheme.colorScheme.tertiaryContainer
),
shape = RoundedCornerShape(16.dp)
) {
Row(
modifier = Modifier
.fillMaxWidth()
.padding(16.dp),
horizontalArrangement = Arrangement.SpaceBetween,
verticalAlignment = Alignment.CenterVertically
) {
Column(modifier = Modifier.weight(1f)) {
Row(
horizontalArrangement = Arrangement.spacedBy(8.dp),
verticalAlignment = Alignment.CenterVertically
) {
Icon(
Icons.Default.AutoFixHigh,
contentDescription = null,
tint = MaterialTheme.colorScheme.tertiary
)
Text(
"Smart Filtering",
style = MaterialTheme.typography.titleMedium,
fontWeight = FontWeight.Bold
)
}
Spacer(Modifier.height(4.dp))
Text(
"Only show photos with detected faces (${faceTaggedUris.size} available)",
style = MaterialTheme.typography.bodySmall,
color = MaterialTheme.colorScheme.onTertiaryContainer.copy(alpha = 0.8f)
)
}
Switch(
checked = onlyShowFaceImages,
onCheckedChange = { onlyShowFaceImages = it }
)
}
}
}
// Gradient header with tips
Card(
modifier = Modifier.fillMaxWidth(),
colors = CardDefaults.cardColors(
containerColor = MaterialTheme.colorScheme.primaryContainer
),
shape = RoundedCornerShape(16.dp)
) {
Column(
modifier = Modifier.padding(20.dp),
verticalArrangement = Arrangement.spacedBy(12.dp)
) {
Row(
horizontalArrangement = Arrangement.spacedBy(12.dp),
verticalAlignment = Alignment.CenterVertically
) {
Surface(
shape = RoundedCornerShape(12.dp),
color = MaterialTheme.colorScheme.primary,
modifier = Modifier.size(48.dp)
) {
Box(contentAlignment = Alignment.Center) {
Icon(
Icons.Default.PhotoCamera,
contentDescription = null,
tint = MaterialTheme.colorScheme.onPrimary,
modifier = Modifier.size(28.dp)
)
}
}
Column {
Text(
"Training Tips",
style = MaterialTheme.typography.titleLarge,
fontWeight = FontWeight.Bold
)
Text(
"More photos = better recognition",
style = MaterialTheme.typography.bodyMedium,
color = MaterialTheme.colorScheme.onPrimaryContainer.copy(alpha = 0.7f)
)
}
}
Spacer(Modifier.height(4.dp))
TipItem("✓ Select 20-30 photos for best results", true)
TipItem("✓ Include different angles and lighting", true)
TipItem("✓ Mix expressions (smile, neutral, laugh)", true)
TipItem("✓ With/without glasses if applicable", true)
TipItem("✗ Avoid blurry or very dark photos", false)
}
}
// Progress indicator
AnimatedVisibility(selectedImages.isNotEmpty()) {
ProgressCard(selectedImages.size)
}
// Select photos button
Button(
onClick = { photoPicker.launch("image/*") },
modifier = Modifier.fillMaxWidth(),
colors = ButtonDefaults.buttonColors(
containerColor = MaterialTheme.colorScheme.primary
),
contentPadding = PaddingValues(vertical = 16.dp)
) {
Icon(Icons.Default.PhotoLibrary, contentDescription = null)
Spacer(Modifier.width(8.dp))
Text(
if (selectedImages.isEmpty()) {
"Select Training Photos"
} else {
"Selected: ${selectedImages.size} photos - Tap to change"
},
style = MaterialTheme.typography.titleMedium
)
}
// Continue button
AnimatedVisibility(selectedImages.size >= 15) {
Button(
onClick = { onImagesSelected(selectedImages) },
modifier = Modifier.fillMaxWidth(),
colors = ButtonDefaults.buttonColors(
containerColor = MaterialTheme.colorScheme.secondary
),
contentPadding = PaddingValues(vertical = 16.dp)
) {
Icon(Icons.Default.Check, contentDescription = null)
Spacer(Modifier.width(8.dp))
Text(
"Continue with ${selectedImages.size} photos",
style = MaterialTheme.typography.titleMedium
)
}
}
// Minimum warning
if (selectedImages.isNotEmpty() && selectedImages.size < 15) {
Card(
modifier = Modifier.fillMaxWidth(),
colors = CardDefaults.cardColors(
containerColor = MaterialTheme.colorScheme.errorContainer
)
) {
Row(
modifier = Modifier.padding(16.dp),
horizontalArrangement = Arrangement.spacedBy(12.dp),
verticalAlignment = Alignment.CenterVertically
) {
Icon(
Icons.Default.Warning,
contentDescription = null,
tint = MaterialTheme.colorScheme.error
)
Column {
Text(
"Need at least 15 photos",
style = MaterialTheme.typography.titleSmall,
fontWeight = FontWeight.Bold,
color = MaterialTheme.colorScheme.onErrorContainer
)
Text(
"You have ${selectedImages.size}. Select ${15 - selectedImages.size} more.",
style = MaterialTheme.typography.bodySmall,
color = MaterialTheme.colorScheme.onErrorContainer.copy(alpha = 0.8f)
)
}
}
}
}
// Bottom spacing to ensure last item is visible
Spacer(Modifier.height(32.dp))
}
}
}
@Composable
private fun TipItem(text: String, isGood: Boolean) {
Row(
horizontalArrangement = Arrangement.spacedBy(8.dp),
verticalAlignment = Alignment.Top
) {
Icon(
if (isGood) Icons.Default.CheckCircle else Icons.Default.Cancel,
contentDescription = null,
modifier = Modifier.size(18.dp),
tint = if (isGood) {
MaterialTheme.colorScheme.primary
} else {
MaterialTheme.colorScheme.error
}
)
Text(
text = text,
style = MaterialTheme.typography.bodyMedium,
color = MaterialTheme.colorScheme.onPrimaryContainer
)
}
}
@Composable
private fun ProgressCard(photoCount: Int) {
Card(
modifier = Modifier.fillMaxWidth(),
colors = CardDefaults.cardColors(
containerColor = when {
photoCount >= 25 -> MaterialTheme.colorScheme.primaryContainer
photoCount >= 20 -> MaterialTheme.colorScheme.tertiaryContainer
else -> MaterialTheme.colorScheme.surfaceVariant
}
),
shape = RoundedCornerShape(16.dp)
) {
Column(
modifier = Modifier.padding(20.dp),
verticalArrangement = Arrangement.spacedBy(12.dp)
) {
Row(
modifier = Modifier.fillMaxWidth(),
horizontalArrangement = Arrangement.SpaceBetween,
verticalAlignment = Alignment.CenterVertically
) {
Column {
Text(
text = "$photoCount photos selected",
style = MaterialTheme.typography.titleMedium,
fontWeight = FontWeight.Bold
)
Text(
text = when {
photoCount >= 30 -> "Excellent! Maximum diversity"
photoCount >= 25 -> "Great! Very good coverage"
photoCount >= 20 -> "Good! Should work well"
photoCount >= 15 -> "Acceptable - more is better"
else -> "Need ${15 - photoCount} more"
},
style = MaterialTheme.typography.bodyMedium,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
}
Surface(
shape = RoundedCornerShape(12.dp),
color = when {
photoCount >= 25 -> MaterialTheme.colorScheme.primary
photoCount >= 20 -> MaterialTheme.colorScheme.tertiary
photoCount >= 15 -> MaterialTheme.colorScheme.secondary
else -> MaterialTheme.colorScheme.outline
},
modifier = Modifier.size(56.dp)
) {
Box(contentAlignment = Alignment.Center) {
Text(
text = when {
photoCount >= 25 -> ""
photoCount >= 20 -> ""
photoCount >= 15 -> ""
else -> "..."
},
style = MaterialTheme.typography.headlineMedium,
color = Color.White
)
}
}
}
// Progress bar
LinearProgressIndicator(
progress = { (photoCount / 30f).coerceAtMost(1f) },
modifier = Modifier
.fillMaxWidth()
.height(8.dp),
color = when {
photoCount >= 25 -> MaterialTheme.colorScheme.primary
photoCount >= 20 -> MaterialTheme.colorScheme.tertiary
else -> MaterialTheme.colorScheme.secondary
},
trackColor = MaterialTheme.colorScheme.surfaceVariant,
)
// Expected accuracy
Row(
modifier = Modifier.fillMaxWidth(),
horizontalArrangement = Arrangement.SpaceBetween
) {
Text(
"Expected accuracy:",
style = MaterialTheme.typography.labelMedium,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
Text(
when {
photoCount >= 30 -> "90-95%"
photoCount >= 25 -> "85-90%"
photoCount >= 20 -> "80-85%"
photoCount >= 15 -> "75-80%"
else -> "< 75%"
},
style = MaterialTheme.typography.labelLarge,
fontWeight = FontWeight.Bold,
color = when {
photoCount >= 25 -> MaterialTheme.colorScheme.primary
photoCount >= 20 -> MaterialTheme.colorScheme.tertiary
else -> MaterialTheme.colorScheme.secondary
}
)
}
}
}
}

View File

@@ -0,0 +1,42 @@
package com.placeholder.sherpai2.ui.trainingprep
import androidx.lifecycle.ViewModel
import androidx.lifecycle.viewModelScope
import com.placeholder.sherpai2.data.local.dao.ImageDao
import dagger.hilt.android.lifecycle.HiltViewModel
import kotlinx.coroutines.flow.MutableStateFlow
import kotlinx.coroutines.flow.StateFlow
import kotlinx.coroutines.flow.asStateFlow
import kotlinx.coroutines.launch
import javax.inject.Inject
/**
* ImageSelectorViewModel
*
* Provides face-tagged image URIs for smart filtering
* during training photo selection
*/
@HiltViewModel
class ImageSelectorViewModel @Inject constructor(
private val imageDao: ImageDao
) : ViewModel() {
private val _faceTaggedImageUris = MutableStateFlow<List<String>>(emptyList())
val faceTaggedImageUris: StateFlow<List<String>> = _faceTaggedImageUris.asStateFlow()
init {
loadFaceTaggedImages()
}
private fun loadFaceTaggedImages() {
viewModelScope.launch {
try {
val imagesWithFaces = imageDao.getImagesWithFaces()
_faceTaggedImageUris.value = imagesWithFaces.map { it.imageUri }
} catch (e: Exception) {
// If cache not available, just use empty list (filter disabled)
_faceTaggedImageUris.value = emptyList()
}
}
}
}

View File

@@ -0,0 +1,231 @@
package com.placeholder.sherpai2.ui.trainingprep
import androidx.compose.foundation.layout.*
import androidx.compose.foundation.shape.RoundedCornerShape
import androidx.compose.foundation.text.KeyboardActions
import androidx.compose.foundation.text.KeyboardOptions
import androidx.compose.material.icons.Icons
import androidx.compose.material.icons.filled.*
import androidx.compose.material3.*
import androidx.compose.runtime.*
import androidx.compose.ui.Alignment
import androidx.compose.ui.Modifier
import androidx.compose.ui.text.font.FontWeight
import androidx.compose.ui.text.input.ImeAction
import androidx.compose.ui.text.input.KeyboardCapitalization
import androidx.compose.ui.unit.dp
import androidx.compose.ui.window.Dialog
/**
* IMPROVED NameInputDialog - Better centered, cleaner layout
*
* Fixes:
* - Centered dialog with proper constraints
* - Better spacing and padding
* - Clearer visual hierarchy
* - Improved error state handling
*/
@OptIn(ExperimentalMaterial3Api::class)
@Composable
fun ImprovedNameInputDialog(
onDismiss: () -> Unit,
onConfirm: (String) -> Unit,
trainingState: TrainingState
) {
var personName by remember { mutableStateOf("") }
val isError = trainingState is TrainingState.Error
val isProcessing = trainingState is TrainingState.Processing
Dialog(
onDismissRequest = {
if (!isProcessing) {
onDismiss()
}
}
) {
Card(
modifier = Modifier
.fillMaxWidth(0.9f) // 90% of screen width
.wrapContentHeight(),
shape = RoundedCornerShape(24.dp),
colors = CardDefaults.cardColors(
containerColor = MaterialTheme.colorScheme.surface
),
elevation = CardDefaults.cardElevation(defaultElevation = 8.dp)
) {
Column(
modifier = Modifier
.fillMaxWidth()
.padding(24.dp),
verticalArrangement = Arrangement.spacedBy(20.dp),
horizontalAlignment = Alignment.CenterHorizontally
) {
// Icon
Surface(
shape = RoundedCornerShape(16.dp),
color = if (isError) {
MaterialTheme.colorScheme.errorContainer
} else {
MaterialTheme.colorScheme.primaryContainer
},
modifier = Modifier.size(72.dp)
) {
Box(contentAlignment = Alignment.Center) {
Icon(
if (isError) Icons.Default.Warning else Icons.Default.Face,
contentDescription = null,
modifier = Modifier.size(40.dp),
tint = if (isError) {
MaterialTheme.colorScheme.error
} else {
MaterialTheme.colorScheme.primary
}
)
}
}
// Title
Text(
text = if (isError) "Training Error" else "Who is this?",
style = MaterialTheme.typography.headlineSmall,
fontWeight = FontWeight.Bold,
color = MaterialTheme.colorScheme.onSurface
)
// Error message or description
if (isError) {
val error = trainingState as TrainingState.Error
Card(
colors = CardDefaults.cardColors(
containerColor = MaterialTheme.colorScheme.errorContainer
),
shape = RoundedCornerShape(12.dp)
) {
Row(
modifier = Modifier
.fillMaxWidth()
.padding(16.dp),
horizontalArrangement = Arrangement.spacedBy(12.dp),
verticalAlignment = Alignment.CenterVertically
) {
Icon(
Icons.Default.Error,
contentDescription = null,
tint = MaterialTheme.colorScheme.error
)
Text(
text = error.message,
style = MaterialTheme.typography.bodyMedium,
color = MaterialTheme.colorScheme.onErrorContainer
)
}
}
} else {
Text(
text = "Enter the name of the person in these training images. This will help you find their photos later.",
style = MaterialTheme.typography.bodyMedium,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
}
// Name input field
OutlinedTextField(
value = personName,
onValueChange = { personName = it },
label = { Text("Person's Name") },
placeholder = { Text("e.g., John Doe") },
singleLine = true,
enabled = !isProcessing,
keyboardOptions = KeyboardOptions(
capitalization = KeyboardCapitalization.Words,
imeAction = ImeAction.Done
),
keyboardActions = KeyboardActions(
onDone = {
if (personName.isNotBlank() && !isProcessing) {
onConfirm(personName.trim())
}
}
),
modifier = Modifier.fillMaxWidth(),
shape = RoundedCornerShape(12.dp),
colors = OutlinedTextFieldDefaults.colors(
focusedBorderColor = MaterialTheme.colorScheme.primary,
unfocusedBorderColor = MaterialTheme.colorScheme.outline
)
)
// Action buttons
Row(
modifier = Modifier.fillMaxWidth(),
horizontalArrangement = Arrangement.spacedBy(12.dp)
) {
// Cancel button
if (!isProcessing) {
OutlinedButton(
onClick = onDismiss,
modifier = Modifier.weight(1f),
shape = RoundedCornerShape(12.dp),
contentPadding = PaddingValues(vertical = 14.dp)
) {
Text("Cancel")
}
}
// Confirm button
Button(
onClick = { onConfirm(personName.trim()) },
enabled = personName.isNotBlank() && !isProcessing,
modifier = Modifier.weight(if (isProcessing) 1f else 1f),
shape = RoundedCornerShape(12.dp),
contentPadding = PaddingValues(vertical = 14.dp),
colors = ButtonDefaults.buttonColors(
containerColor = if (isError) {
MaterialTheme.colorScheme.error
} else {
MaterialTheme.colorScheme.primary
}
)
) {
if (isProcessing) {
CircularProgressIndicator(
modifier = Modifier.size(20.dp),
strokeWidth = 2.dp,
color = MaterialTheme.colorScheme.onPrimary
)
Spacer(modifier = Modifier.width(12.dp))
Text("Training...")
} else {
Icon(
if (isError) Icons.Default.Refresh else Icons.Default.Check,
contentDescription = null,
modifier = Modifier.size(20.dp)
)
Spacer(modifier = Modifier.width(8.dp))
Text(if (isError) "Try Again" else "Start Training")
}
}
}
}
}
}
}
/**
* Alternative: Use this version in ScanResultsScreen.kt
*
* Replace the existing NameInputDialog function (lines 154-257) with:
*
* @Composable
* private fun NameInputDialog(
* onDismiss: () -> Unit,
* onConfirm: (String) -> Unit,
* trainingState: TrainingState
* ) {
* ImprovedNameInputDialog(
* onDismiss = onDismiss,
* onConfirm = onConfirm,
* trainingState = trainingState
* )
* }
*/

View File

@@ -0,0 +1,547 @@
package com.placeholder.sherpai2.ui.trainingprep
import android.net.Uri
import androidx.activity.compose.rememberLauncherForActivityResult
import androidx.activity.result.PickVisualMediaRequest
import androidx.activity.result.contract.ActivityResultContracts
import androidx.compose.foundation.BorderStroke
import androidx.compose.foundation.Image
import androidx.compose.foundation.background
import androidx.compose.foundation.border
import androidx.compose.foundation.layout.*
import androidx.compose.foundation.lazy.LazyColumn
import androidx.compose.foundation.lazy.itemsIndexed
import androidx.compose.foundation.shape.CircleShape
import androidx.compose.foundation.shape.RoundedCornerShape
import androidx.compose.material.icons.Icons
import androidx.compose.material.icons.filled.*
import androidx.compose.material3.*
import androidx.compose.runtime.*
import androidx.compose.ui.Alignment
import androidx.compose.ui.Modifier
import androidx.compose.ui.draw.clip
import androidx.compose.ui.graphics.Color
import androidx.compose.ui.graphics.asImageBitmap
import androidx.compose.ui.layout.ContentScale
import androidx.compose.ui.text.font.FontWeight
import androidx.compose.ui.text.style.TextAlign
import androidx.compose.ui.unit.dp
import androidx.hilt.navigation.compose.hiltViewModel
import coil.compose.AsyncImage
@OptIn(ExperimentalMaterial3Api::class)
@Composable
fun ScanResultsScreen(
state: ScanningState,
onFinish: () -> Unit,
trainViewModel: TrainViewModel = hiltViewModel()
) {
var showFacePickerDialog by remember { mutableStateOf<FaceDetectionHelper.FaceDetectionResult?>(null) }
val trainingState by trainViewModel.trainingState.collectAsState()
LaunchedEffect(trainingState) {
when (trainingState) {
is TrainingState.Success -> {
trainViewModel.resetTrainingState()
onFinish()
}
is TrainingState.Error -> {}
else -> {}
}
}
Scaffold(
topBar = {
TopAppBar(
title = { Text("Train New Person") },
colors = TopAppBarDefaults.topAppBarColors(
containerColor = MaterialTheme.colorScheme.primaryContainer
)
)
}
) { paddingValues ->
Box(
modifier = Modifier
.fillMaxSize()
.padding(paddingValues)
) {
when (state) {
is ScanningState.Idle -> {}
is ScanningState.Processing -> {
ProcessingView(progress = state.progress, total = state.total)
}
is ScanningState.Success -> {
ImprovedResultsView(
result = state.sanityCheckResult,
onContinue = {
// PersonInfo already captured in TrainingScreen!
// Just start training with stored info
trainViewModel.createFaceModel(
trainViewModel.getPersonInfo()?.name ?: "Unknown"
)
},
onRetry = onFinish,
onReplaceImage = { oldUri, newUri ->
trainViewModel.replaceImage(oldUri, newUri)
},
onSelectFaceFromMultiple = { result ->
showFacePickerDialog = result
},
trainViewModel = trainViewModel
)
}
is ScanningState.Error -> {
ErrorView(message = state.message, onRetry = onFinish)
}
}
if (trainingState is TrainingState.Processing) {
TrainingOverlay(trainingState = trainingState as TrainingState.Processing)
}
}
}
showFacePickerDialog?.let { result ->
FacePickerDialog(
result = result,
onDismiss = { showFacePickerDialog = null },
onFaceSelected = { faceIndex, croppedFaceBitmap ->
trainViewModel.selectFaceFromImage(result.uri, faceIndex, croppedFaceBitmap)
showFacePickerDialog = null
}
)
}
}
@Composable
private fun TrainingOverlay(trainingState: TrainingState.Processing) {
Box(
modifier = Modifier.fillMaxSize().background(Color.Black.copy(alpha = 0.7f)),
contentAlignment = Alignment.Center
) {
Card(
modifier = Modifier.padding(32.dp).fillMaxWidth(0.9f),
colors = CardDefaults.cardColors(containerColor = MaterialTheme.colorScheme.surface)
) {
Column(
modifier = Modifier.padding(24.dp),
horizontalAlignment = Alignment.CenterHorizontally,
verticalArrangement = Arrangement.spacedBy(16.dp)
) {
CircularProgressIndicator(modifier = Modifier.size(64.dp), strokeWidth = 6.dp)
Text("Creating Face Model", style = MaterialTheme.typography.titleLarge, fontWeight = FontWeight.Bold)
Text(trainingState.stage, style = MaterialTheme.typography.bodyMedium, textAlign = TextAlign.Center, color = MaterialTheme.colorScheme.onSurfaceVariant)
if (trainingState.total > 0) {
LinearProgressIndicator(
progress = { (trainingState.progress.toFloat() / trainingState.total.toFloat()).coerceIn(0f, 1f) },
modifier = Modifier.fillMaxWidth()
)
Text("${trainingState.progress} / ${trainingState.total}", style = MaterialTheme.typography.bodySmall, color = MaterialTheme.colorScheme.onSurfaceVariant)
}
}
}
}
}
@Composable
private fun ProcessingView(progress: Int, total: Int) {
Column(
modifier = Modifier.fillMaxSize(),
horizontalAlignment = Alignment.CenterHorizontally,
verticalArrangement = Arrangement.Center
) {
CircularProgressIndicator(modifier = Modifier.size(64.dp), strokeWidth = 6.dp)
Spacer(modifier = Modifier.height(24.dp))
Text("Analyzing images...", style = MaterialTheme.typography.titleMedium)
Spacer(modifier = Modifier.height(8.dp))
Text("Detecting faces and checking for duplicates", style = MaterialTheme.typography.bodyMedium, color = MaterialTheme.colorScheme.onSurfaceVariant)
if (total > 0) {
Spacer(modifier = Modifier.height(16.dp))
LinearProgressIndicator(
progress = { (progress.toFloat() / total.toFloat()).coerceIn(0f, 1f) },
modifier = Modifier.width(200.dp)
)
Text("$progress / $total", style = MaterialTheme.typography.bodySmall)
}
}
}
@Composable
private fun ImprovedResultsView(
result: TrainingSanityChecker.SanityCheckResult,
onContinue: () -> Unit,
onRetry: () -> Unit,
onReplaceImage: (Uri, Uri) -> Unit,
onSelectFaceFromMultiple: (FaceDetectionHelper.FaceDetectionResult) -> Unit,
trainViewModel: TrainViewModel
) {
LazyColumn(
modifier = Modifier.fillMaxSize(),
contentPadding = PaddingValues(16.dp),
verticalArrangement = Arrangement.spacedBy(16.dp)
) {
item {
Card(
modifier = Modifier.fillMaxWidth(),
colors = CardDefaults.cardColors(containerColor = MaterialTheme.colorScheme.secondaryContainer)
) {
Column(modifier = Modifier.padding(16.dp)) {
Text("Analysis Complete!", style = MaterialTheme.typography.headlineSmall, fontWeight = FontWeight.Bold)
Spacer(modifier = Modifier.height(4.dp))
Text(
"Review your images below. Tap 'Pick Face' on group photos to choose which person to train on, or 'Replace' to swap out any image.",
style = MaterialTheme.typography.bodyMedium,
color = MaterialTheme.colorScheme.onSecondaryContainer.copy(alpha = 0.8f)
)
}
}
}
item {
ProgressSummaryCard(
totalImages = result.faceDetectionResults.size,
validImages = result.validImagesWithFaces.size,
requiredImages = 10,
isValid = result.isValid
)
}
item {
Text("Your Images (${result.faceDetectionResults.size})", style = MaterialTheme.typography.titleLarge, fontWeight = FontWeight.Bold)
}
itemsIndexed(result.faceDetectionResults) { index, imageResult ->
ImageResultCard(
index = index + 1,
result = imageResult,
onReplace = { newUri -> onReplaceImage(imageResult.uri, newUri) },
onSelectFace = if (imageResult.faceCount > 1) { { onSelectFaceFromMultiple(imageResult) } } else null,
trainViewModel = trainViewModel,
isExcluded = trainViewModel.isImageExcluded(imageResult.uri)
)
}
if (result.validationErrors.isNotEmpty()) {
item {
Spacer(modifier = Modifier.height(8.dp))
ValidationIssuesCard(errors = result.validationErrors, trainViewModel = trainViewModel)
}
}
item {
Spacer(modifier = Modifier.height(8.dp))
Button(
onClick = if (result.isValid) onContinue else onRetry,
modifier = Modifier.fillMaxWidth(),
enabled = result.isValid,
colors = ButtonDefaults.buttonColors(
containerColor = if (result.isValid) MaterialTheme.colorScheme.primary else MaterialTheme.colorScheme.error.copy(alpha = 0.5f)
)
) {
Icon(if (result.isValid) Icons.Default.CheckCircle else Icons.Default.Warning, contentDescription = null)
Spacer(modifier = Modifier.width(8.dp))
Text(
if (result.isValid)
"Continue to Training (${result.validImagesWithFaces.size} images)"
else
"Fix ${result.validationErrors.size} Issue${if (result.validationErrors.size != 1) "s" else ""} to Continue"
)
}
if (!result.isValid) {
Spacer(modifier = Modifier.height(8.dp))
Surface(
modifier = Modifier.fillMaxWidth(),
color = MaterialTheme.colorScheme.tertiaryContainer,
shape = RoundedCornerShape(8.dp)
) {
Row(modifier = Modifier.padding(12.dp), verticalAlignment = Alignment.CenterVertically) {
Icon(Icons.Default.Info, contentDescription = null, tint = MaterialTheme.colorScheme.onTertiaryContainer, modifier = Modifier.size(20.dp))
Spacer(modifier = Modifier.width(8.dp))
Text(
"Tip: Use 'Replace' to swap problematic images, or 'Pick Face' to choose from group photos",
style = MaterialTheme.typography.bodySmall,
color = MaterialTheme.colorScheme.onTertiaryContainer
)
}
}
}
}
}
}
@Composable
private fun ProgressSummaryCard(totalImages: Int, validImages: Int, requiredImages: Int, isValid: Boolean) {
Card(
modifier = Modifier.fillMaxWidth(),
colors = CardDefaults.cardColors(
containerColor = if (isValid) MaterialTheme.colorScheme.primaryContainer.copy(alpha = 0.5f) else MaterialTheme.colorScheme.errorContainer.copy(alpha = 0.3f)
)
) {
Column(modifier = Modifier.padding(16.dp)) {
Row(modifier = Modifier.fillMaxWidth(), horizontalArrangement = Arrangement.SpaceBetween, verticalAlignment = Alignment.CenterVertically) {
Text("Progress", style = MaterialTheme.typography.titleMedium, fontWeight = FontWeight.Bold)
Icon(
imageVector = if (isValid) Icons.Default.CheckCircle else Icons.Default.Warning,
contentDescription = null,
tint = if (isValid) MaterialTheme.colorScheme.primary else MaterialTheme.colorScheme.error,
modifier = Modifier.size(32.dp)
)
}
Spacer(modifier = Modifier.height(12.dp))
Row(modifier = Modifier.fillMaxWidth(), horizontalArrangement = Arrangement.SpaceEvenly) {
StatItem("Total", totalImages.toString(), MaterialTheme.colorScheme.onSurface)
StatItem("Valid", validImages.toString(), if (validImages >= requiredImages) MaterialTheme.colorScheme.primary else MaterialTheme.colorScheme.error)
StatItem("Need", requiredImages.toString(), MaterialTheme.colorScheme.onSurface.copy(alpha = 0.6f))
}
Spacer(modifier = Modifier.height(12.dp))
LinearProgressIndicator(
progress = { (validImages.toFloat() / requiredImages.toFloat()).coerceIn(0f, 1f) },
modifier = Modifier.fillMaxWidth(),
color = if (isValid) MaterialTheme.colorScheme.primary else MaterialTheme.colorScheme.error
)
}
}
}
@Composable
private fun StatItem(label: String, value: String, color: Color) {
Column(horizontalAlignment = Alignment.CenterHorizontally) {
Text(value, style = MaterialTheme.typography.headlineMedium, fontWeight = FontWeight.Bold, color = color)
Text(label, style = MaterialTheme.typography.bodySmall, color = color.copy(alpha = 0.7f))
}
}
@Composable
private fun ImageResultCard(
index: Int,
result: FaceDetectionHelper.FaceDetectionResult,
onReplace: (Uri) -> Unit,
onSelectFace: (() -> Unit)?,
trainViewModel: TrainViewModel,
isExcluded: Boolean
) {
val photoPickerLauncher = rememberLauncherForActivityResult(contract = ActivityResultContracts.PickVisualMedia()) { uri -> uri?.let { onReplace(it) } }
val status = when {
isExcluded -> ImageStatus.EXCLUDED
result.errorMessage != null -> ImageStatus.ERROR
!result.hasFace -> ImageStatus.NO_FACE
result.faceCount > 1 -> ImageStatus.MULTIPLE_FACES
result.faceCount == 1 -> ImageStatus.VALID
else -> ImageStatus.ERROR
}
Card(
modifier = Modifier.fillMaxWidth(),
colors = CardDefaults.cardColors(
containerColor = when (status) {
ImageStatus.VALID -> MaterialTheme.colorScheme.primaryContainer.copy(alpha = 0.3f)
ImageStatus.MULTIPLE_FACES -> MaterialTheme.colorScheme.tertiaryContainer.copy(alpha = 0.4f)
ImageStatus.EXCLUDED -> MaterialTheme.colorScheme.surfaceVariant.copy(alpha = 0.5f)
else -> MaterialTheme.colorScheme.errorContainer.copy(alpha = 0.3f)
}
)
) {
Row(modifier = Modifier.fillMaxWidth().padding(12.dp), verticalAlignment = Alignment.CenterVertically, horizontalArrangement = Arrangement.spacedBy(12.dp)) {
Box(
modifier = Modifier.size(40.dp).background(
color = when (status) {
ImageStatus.VALID -> MaterialTheme.colorScheme.primary
ImageStatus.MULTIPLE_FACES -> MaterialTheme.colorScheme.tertiary
ImageStatus.EXCLUDED -> MaterialTheme.colorScheme.outline
else -> MaterialTheme.colorScheme.error
},
shape = CircleShape
),
contentAlignment = Alignment.Center
) {
Text(index.toString(), style = MaterialTheme.typography.titleMedium, fontWeight = FontWeight.Bold, color = Color.White)
}
if (result.croppedFaceBitmap != null) {
Image(
bitmap = result.croppedFaceBitmap.asImageBitmap(),
contentDescription = "Face",
modifier = Modifier.size(64.dp).clip(RoundedCornerShape(8.dp)).border(
BorderStroke(2.dp, when (status) {
ImageStatus.VALID -> MaterialTheme.colorScheme.primary
ImageStatus.MULTIPLE_FACES -> MaterialTheme.colorScheme.tertiary
ImageStatus.EXCLUDED -> MaterialTheme.colorScheme.outline
else -> MaterialTheme.colorScheme.error
}),
RoundedCornerShape(8.dp)
),
contentScale = ContentScale.Crop
)
} else {
AsyncImage(model = result.uri, contentDescription = "Original image", modifier = Modifier.size(64.dp).clip(RoundedCornerShape(8.dp)), contentScale = ContentScale.Crop)
}
Column(modifier = Modifier.weight(1f)) {
Row(verticalAlignment = Alignment.CenterVertically) {
Icon(
imageVector = when (status) {
ImageStatus.VALID -> Icons.Default.CheckCircle
ImageStatus.MULTIPLE_FACES -> Icons.Default.Info
ImageStatus.EXCLUDED -> Icons.Default.RemoveCircle
else -> Icons.Default.Warning
},
contentDescription = null,
tint = when (status) {
ImageStatus.VALID -> MaterialTheme.colorScheme.primary
ImageStatus.MULTIPLE_FACES -> MaterialTheme.colorScheme.tertiary
ImageStatus.EXCLUDED -> MaterialTheme.colorScheme.outline
else -> MaterialTheme.colorScheme.error
},
modifier = Modifier.size(20.dp)
)
Spacer(modifier = Modifier.width(4.dp))
Text(
text = when (status) {
ImageStatus.VALID -> "Face Detected"
ImageStatus.MULTIPLE_FACES -> "Multiple Faces (${result.faceCount})"
ImageStatus.NO_FACE -> "No Face Detected"
ImageStatus.EXCLUDED -> "Excluded"
ImageStatus.ERROR -> "Error"
},
style = MaterialTheme.typography.bodyMedium,
fontWeight = FontWeight.SemiBold
)
}
Text(result.uri.lastPathSegment ?: "Unknown", style = MaterialTheme.typography.bodySmall, color = MaterialTheme.colorScheme.onSurfaceVariant, maxLines = 1)
}
Column(horizontalAlignment = Alignment.End, verticalArrangement = Arrangement.spacedBy(4.dp)) {
if (onSelectFace != null && !isExcluded) {
OutlinedButton(
onClick = onSelectFace,
modifier = Modifier.height(32.dp),
contentPadding = PaddingValues(horizontal = 12.dp, vertical = 0.dp),
colors = ButtonDefaults.outlinedButtonColors(contentColor = MaterialTheme.colorScheme.tertiary),
border = BorderStroke(1.dp, MaterialTheme.colorScheme.tertiary)
) {
Icon(Icons.Default.Face, contentDescription = null, modifier = Modifier.size(16.dp))
Spacer(modifier = Modifier.width(4.dp))
Text("Pick Face", style = MaterialTheme.typography.bodySmall)
}
}
if (!isExcluded) {
OutlinedButton(
onClick = { photoPickerLauncher.launch(PickVisualMediaRequest(ActivityResultContracts.PickVisualMedia.ImageOnly)) },
modifier = Modifier.height(32.dp),
contentPadding = PaddingValues(horizontal = 12.dp, vertical = 0.dp)
) {
Icon(Icons.Default.Refresh, contentDescription = null, modifier = Modifier.size(16.dp))
Spacer(modifier = Modifier.width(4.dp))
Text("Replace", style = MaterialTheme.typography.bodySmall)
}
}
OutlinedButton(
onClick = {
if (isExcluded) trainViewModel.includeImage(result.uri) else trainViewModel.excludeImage(result.uri)
},
modifier = Modifier.height(32.dp),
contentPadding = PaddingValues(horizontal = 12.dp, vertical = 0.dp),
colors = ButtonDefaults.outlinedButtonColors(contentColor = if (isExcluded) MaterialTheme.colorScheme.primary else MaterialTheme.colorScheme.error),
border = BorderStroke(1.dp, if (isExcluded) MaterialTheme.colorScheme.primary else MaterialTheme.colorScheme.error)
) {
Icon(if (isExcluded) Icons.Default.Add else Icons.Default.Close, contentDescription = null, modifier = Modifier.size(16.dp))
Spacer(modifier = Modifier.width(4.dp))
Text(if (isExcluded) "Include" else "Exclude", style = MaterialTheme.typography.bodySmall)
}
}
}
}
}
@Composable
private fun ValidationIssuesCard(errors: List<TrainingSanityChecker.ValidationError>, trainViewModel: TrainViewModel) {
Card(
modifier = Modifier.fillMaxWidth(),
colors = CardDefaults.cardColors(containerColor = MaterialTheme.colorScheme.errorContainer.copy(alpha = 0.3f))
) {
Column(modifier = Modifier.padding(16.dp), verticalArrangement = Arrangement.spacedBy(8.dp)) {
Row(verticalAlignment = Alignment.CenterVertically) {
Icon(Icons.Default.Warning, contentDescription = null, tint = MaterialTheme.colorScheme.error)
Spacer(modifier = Modifier.width(8.dp))
Text("Issues Found (${errors.size})", style = MaterialTheme.typography.titleMedium, fontWeight = FontWeight.Bold, color = MaterialTheme.colorScheme.error)
}
HorizontalDivider(color = MaterialTheme.colorScheme.error.copy(alpha = 0.3f))
errors.forEach { error ->
when (error) {
is TrainingSanityChecker.ValidationError.NoFaceDetected -> {
Text("${error.uris.size} image(s) without detected faces - use Replace button", style = MaterialTheme.typography.bodyMedium)
}
is TrainingSanityChecker.ValidationError.MultipleFacesDetected -> {
Text("${error.uri.lastPathSegment} has ${error.faceCount} faces - use Pick Face button", style = MaterialTheme.typography.bodyMedium)
}
is TrainingSanityChecker.ValidationError.DuplicateImages -> {
Column(verticalArrangement = Arrangement.spacedBy(8.dp)) {
Row(modifier = Modifier.fillMaxWidth(), horizontalArrangement = Arrangement.SpaceBetween, verticalAlignment = Alignment.CenterVertically) {
Text("${error.groups.size} duplicate group(s) found", style = MaterialTheme.typography.bodyMedium, modifier = Modifier.weight(1f))
Button(
onClick = {
error.groups.forEach { group ->
group.images.drop(1).forEach { uri ->
trainViewModel.excludeImage(uri)
}
}
},
colors = ButtonDefaults.buttonColors(containerColor = MaterialTheme.colorScheme.tertiary),
modifier = Modifier.height(36.dp)
) {
Icon(Icons.Default.DeleteSweep, contentDescription = null, modifier = Modifier.size(16.dp))
Spacer(Modifier.width(4.dp))
Text("Drop All", style = MaterialTheme.typography.labelMedium)
}
}
Text("${error.groups.sumOf { it.images.size - 1 }} duplicate images will be excluded", style = MaterialTheme.typography.bodySmall, color = MaterialTheme.colorScheme.onSurface.copy(alpha = 0.6f))
}
}
is TrainingSanityChecker.ValidationError.InsufficientImages -> {
Text("• Need ${error.required} valid images, currently have ${error.available}", style = MaterialTheme.typography.bodyMedium, fontWeight = FontWeight.Bold)
}
is TrainingSanityChecker.ValidationError.ImageLoadError -> {
Text("• Failed to load ${error.uri.lastPathSegment} - use Replace button", style = MaterialTheme.typography.bodyMedium)
}
}
}
}
}
}
@Composable
private fun ErrorView(message: String, onRetry: () -> Unit) {
Column(modifier = Modifier.fillMaxSize().padding(16.dp), horizontalAlignment = Alignment.CenterHorizontally, verticalArrangement = Arrangement.Center) {
Icon(imageVector = Icons.Default.Close, contentDescription = null, modifier = Modifier.size(64.dp), tint = MaterialTheme.colorScheme.error)
Spacer(modifier = Modifier.height(16.dp))
Text("Error", style = MaterialTheme.typography.titleLarge, fontWeight = FontWeight.Bold)
Spacer(modifier = Modifier.height(8.dp))
Text(message, style = MaterialTheme.typography.bodyMedium, textAlign = TextAlign.Center)
Spacer(modifier = Modifier.height(24.dp))
Button(onClick = onRetry) {
Icon(Icons.Default.Refresh, contentDescription = null)
Spacer(modifier = Modifier.width(8.dp))
Text("Try Again")
}
}
}
private enum class ImageStatus {
VALID,
MULTIPLE_FACES,
NO_FACE,
ERROR,
EXCLUDED
}

View File

@@ -0,0 +1,418 @@
package com.placeholder.sherpai2.ui.trainingprep
import android.app.Application
import android.graphics.Bitmap
import android.net.Uri
import androidx.lifecycle.AndroidViewModel
import androidx.lifecycle.viewModelScope
import com.placeholder.sherpai2.data.local.entity.PersonEntity
import com.placeholder.sherpai2.data.repository.FaceRecognitionRepository
import com.placeholder.sherpai2.ml.FaceNetModel
import dagger.hilt.android.lifecycle.HiltViewModel
import kotlinx.coroutines.flow.MutableStateFlow
import kotlinx.coroutines.flow.StateFlow
import kotlinx.coroutines.flow.asStateFlow
import kotlinx.coroutines.launch
import javax.inject.Inject
sealed class ScanningState {
object Idle : ScanningState()
data class Processing(val progress: Int, val total: Int) : ScanningState()
data class Success(
val sanityCheckResult: TrainingSanityChecker.SanityCheckResult
) : ScanningState()
data class Error(val message: String) : ScanningState()
}
sealed class TrainingState {
object Idle : TrainingState()
data class Processing(val stage: String, val progress: Int, val total: Int) : TrainingState()
data class Success(
val personName: String,
val personId: String,
val relationship: String?
) : TrainingState()
data class Error(val message: String) : TrainingState()
}
/**
* Person info captured before photo selection
*/
data class PersonInfo(
val name: String,
val dateOfBirth: Long?,
val relationship: String
)
/**
* FIXED TrainViewModel with proper exclude functionality and efficient replace
*/
@HiltViewModel
class TrainViewModel @Inject constructor(
application: Application,
private val faceRecognitionRepository: FaceRecognitionRepository,
private val faceNetModel: FaceNetModel
) : AndroidViewModel(application) {
private val sanityChecker = TrainingSanityChecker(application)
private val faceDetectionHelper = FaceDetectionHelper(application)
private val _uiState = MutableStateFlow<ScanningState>(ScanningState.Idle)
val uiState: StateFlow<ScanningState> = _uiState.asStateFlow()
private val _trainingState = MutableStateFlow<TrainingState>(TrainingState.Idle)
val trainingState: StateFlow<TrainingState> = _trainingState.asStateFlow()
// Store person info for later use during training
private var personInfo: PersonInfo? = null
private var currentImageUris: List<Uri> = emptyList()
private val manualFaceSelections = mutableMapOf<Uri, ManualFaceSelection>()
// Track excluded images
private val excludedImages = mutableSetOf<Uri>()
data class ManualFaceSelection(
val faceIndex: Int,
val croppedFaceBitmap: Bitmap
)
/**
* Store person info before photo selection
*/
fun setPersonInfo(name: String, dateOfBirth: Long?, relationship: String) {
personInfo = PersonInfo(name, dateOfBirth, relationship)
}
/**
* Get stored person info
*/
fun getPersonInfo(): PersonInfo? = personInfo
/**
* Exclude an image from training
*/
fun excludeImage(uri: Uri) {
excludedImages.add(uri)
val currentState = _uiState.value
if (currentState is ScanningState.Success) {
val updatedResult = applyManualSelections(currentState.sanityCheckResult)
_uiState.value = ScanningState.Success(updatedResult)
}
}
/**
* Include a previously excluded image
*/
fun includeImage(uri: Uri) {
excludedImages.remove(uri)
val currentState = _uiState.value
if (currentState is ScanningState.Success) {
val updatedResult = applyManualSelections(currentState.sanityCheckResult)
_uiState.value = ScanningState.Success(updatedResult)
}
}
/**
* Check if an image is excluded
*/
fun isImageExcluded(uri: Uri): Boolean {
return uri in excludedImages
}
/**
* Create face model with captured person info
*/
fun createFaceModel(personName: String) {
val currentState = _uiState.value
if (currentState !is ScanningState.Success) {
_trainingState.value = TrainingState.Error("No validated images available")
return
}
val validImages = currentState.sanityCheckResult.validImagesWithFaces
if (validImages.size < 15) {
_trainingState.value = TrainingState.Error(
"Need at least 15 valid images, have ${validImages.size}"
)
return
}
viewModelScope.launch {
try {
_trainingState.value = TrainingState.Processing(
stage = "Creating person and training model",
progress = 0,
total = validImages.size
)
val person = PersonEntity.create(
name = personName,
dateOfBirth = personInfo?.dateOfBirth,
relationship = personInfo?.relationship
)
val personId = faceRecognitionRepository.createPersonWithFaceModel(
person = person,
validImages = validImages,
onProgress = { current, total ->
_trainingState.value = TrainingState.Processing(
stage = "Processing image $current/$total",
progress = current,
total = total
)
}
)
_trainingState.value = TrainingState.Success(
personName = personName,
personId = personId,
relationship = person.relationship
)
} catch (e: Exception) {
_trainingState.value = TrainingState.Error(
e.message ?: "Failed to create face model"
)
}
}
}
fun resetTrainingState() {
_trainingState.value = TrainingState.Idle
}
fun scanAndTagFaces(imageUris: List<Uri>) {
currentImageUris = imageUris
manualFaceSelections.clear()
excludedImages.clear()
performScan(imageUris)
}
/**
* FIXED: Replace image - only rescan the ONE new image, not all images!
*/
fun replaceImage(oldUri: Uri, newUri: Uri) {
viewModelScope.launch {
try {
val currentState = _uiState.value
if (currentState !is ScanningState.Success) return@launch
// Update the URI list
val updatedUris = currentImageUris.toMutableList()
val index = updatedUris.indexOf(oldUri)
if (index == -1) return@launch
updatedUris[index] = newUri
currentImageUris = updatedUris
// Clean up old selections/exclusions
manualFaceSelections.remove(oldUri)
excludedImages.remove(oldUri)
// Only scan the NEW image
val newResult = faceDetectionHelper.detectFacesInImage(newUri)
// Update the results list
val updatedFaceResults = currentState.sanityCheckResult.faceDetectionResults.toMutableList()
updatedFaceResults[index] = newResult
// Create updated SanityCheckResult
val updatedSanityResult = currentState.sanityCheckResult.copy(
faceDetectionResults = updatedFaceResults
)
// Apply manual selections and exclusions
val finalResult = applyManualSelections(updatedSanityResult)
_uiState.value = ScanningState.Success(finalResult)
} catch (e: Exception) {
_uiState.value = ScanningState.Error(
e.message ?: "Failed to replace image"
)
}
}
}
/**
* Select face and auto-include the image
*/
fun selectFaceFromImage(imageUri: Uri, faceIndex: Int, croppedFaceBitmap: Bitmap) {
manualFaceSelections[imageUri] = ManualFaceSelection(faceIndex, croppedFaceBitmap)
excludedImages.remove(imageUri) // Auto-include
val currentState = _uiState.value
if (currentState is ScanningState.Success) {
val updatedResult = applyManualSelections(currentState.sanityCheckResult)
_uiState.value = ScanningState.Success(updatedResult)
}
}
/**
* Perform full scan with exclusions and progress tracking
*/
private fun performScan(imageUris: List<Uri>) {
viewModelScope.launch {
try {
_uiState.value = ScanningState.Processing(0, imageUris.size)
val result = sanityChecker.performSanityChecks(
imageUris = imageUris,
minImagesRequired = 15,
allowMultipleFaces = true,
duplicateSimilarityThreshold = 0.95,
excludedImages = excludedImages,
onProgress = { stage, current, total ->
_uiState.value = ScanningState.Processing(current, total)
}
)
val finalResult = applyManualSelections(result)
_uiState.value = ScanningState.Success(finalResult)
} catch (e: Exception) {
_uiState.value = ScanningState.Error(
e.message ?: "An unknown error occurred"
)
}
}
}
/**
* Apply manual selections with exclusion filtering
*/
private fun applyManualSelections(
result: TrainingSanityChecker.SanityCheckResult
): TrainingSanityChecker.SanityCheckResult {
if (manualFaceSelections.isEmpty() && excludedImages.isEmpty()) {
return result
}
val updatedFaceResults = result.faceDetectionResults.map { faceResult ->
val manualSelection = manualFaceSelections[faceResult.uri]
if (manualSelection != null) {
faceResult.copy(
croppedFaceBitmap = manualSelection.croppedFaceBitmap,
faceCount = 1
)
} else {
faceResult
}
}
val updatedValidImages = updatedFaceResults
.filter { it.uri !in excludedImages } // Filter excluded
.filter { it.hasFace }
.filter { it.croppedFaceBitmap != null }
.filter { it.errorMessage == null }
.filter { it.faceCount >= 1 }
.map { faceResult ->
TrainingSanityChecker.ValidTrainingImage(
uri = faceResult.uri,
croppedFaceBitmap = faceResult.croppedFaceBitmap!!,
faceCount = faceResult.faceCount
)
}
val updatedErrors = result.validationErrors.toMutableList()
// Remove errors for manually selected faces or excluded images
updatedErrors.removeAll { error ->
when (error) {
is TrainingSanityChecker.ValidationError.MultipleFacesDetected ->
manualFaceSelections.containsKey(error.uri) || excludedImages.contains(error.uri)
is TrainingSanityChecker.ValidationError.NoFaceDetected ->
error.uris.any { excludedImages.contains(it) }
is TrainingSanityChecker.ValidationError.ImageLoadError ->
excludedImages.contains(error.uri)
else -> false
}
}
// Update insufficient images error
if (updatedValidImages.size < 15) {
if (updatedErrors.none { it is TrainingSanityChecker.ValidationError.InsufficientImages }) {
updatedErrors.add(
TrainingSanityChecker.ValidationError.InsufficientImages(
required = 15,
available = updatedValidImages.size
)
)
}
} else {
updatedErrors.removeAll { it is TrainingSanityChecker.ValidationError.InsufficientImages }
}
val isValid = updatedErrors.isEmpty() && updatedValidImages.size >= 15
return result.copy(
isValid = isValid,
faceDetectionResults = updatedFaceResults,
validationErrors = updatedErrors,
validImagesWithFaces = updatedValidImages,
excludedImages = excludedImages
)
}
fun getFormattedErrors(result: TrainingSanityChecker.SanityCheckResult): List<String> {
return sanityChecker.formatValidationErrors(result.validationErrors)
}
fun reset() {
_uiState.value = ScanningState.Idle
_trainingState.value = TrainingState.Idle
currentImageUris = emptyList()
manualFaceSelections.clear()
excludedImages.clear()
personInfo = null
}
override fun onCleared() {
super.onCleared()
sanityChecker.cleanup()
faceDetectionHelper.cleanup()
faceNetModel.close()
}
}
// Extension functions for copying results
private fun FaceDetectionHelper.FaceDetectionResult.copy(
uri: Uri = this.uri,
hasFace: Boolean = this.hasFace,
faceCount: Int = this.faceCount,
faceBounds: List<android.graphics.Rect> = this.faceBounds,
croppedFaceBitmap: Bitmap? = this.croppedFaceBitmap,
errorMessage: String? = this.errorMessage
): FaceDetectionHelper.FaceDetectionResult {
return FaceDetectionHelper.FaceDetectionResult(
uri = uri,
hasFace = hasFace,
faceCount = faceCount,
faceBounds = faceBounds,
croppedFaceBitmap = croppedFaceBitmap,
errorMessage = errorMessage
)
}
private fun TrainingSanityChecker.SanityCheckResult.copy(
isValid: Boolean = this.isValid,
faceDetectionResults: List<FaceDetectionHelper.FaceDetectionResult> = this.faceDetectionResults,
duplicateCheckResult: DuplicateImageDetector.DuplicateCheckResult = this.duplicateCheckResult,
validationErrors: List<TrainingSanityChecker.ValidationError> = this.validationErrors,
warnings: List<String> = this.warnings,
validImagesWithFaces: List<TrainingSanityChecker.ValidTrainingImage> = this.validImagesWithFaces,
excludedImages: Set<Uri> = this.excludedImages
): TrainingSanityChecker.SanityCheckResult {
return TrainingSanityChecker.SanityCheckResult(
isValid = isValid,
faceDetectionResults = faceDetectionResults,
duplicateCheckResult = duplicateCheckResult,
validationErrors = validationErrors,
warnings = warnings,
validImagesWithFaces = validImagesWithFaces,
excludedImages = excludedImages
)
}

View File

@@ -0,0 +1,205 @@
package com.placeholder.sherpai2.ui.trainingprep
import androidx.compose.foundation.background
import androidx.compose.foundation.layout.*
import androidx.compose.foundation.rememberScrollState
import androidx.compose.foundation.shape.RoundedCornerShape
import androidx.compose.foundation.verticalScroll
import androidx.compose.material.icons.Icons
import androidx.compose.material.icons.filled.*
import androidx.compose.material3.*
import androidx.compose.runtime.*
import androidx.compose.ui.Alignment
import androidx.compose.ui.Modifier
import androidx.compose.ui.graphics.Brush
import androidx.compose.ui.graphics.vector.ImageVector
import androidx.compose.ui.text.font.FontWeight
import androidx.compose.ui.text.style.TextAlign
import androidx.compose.ui.unit.dp
import androidx.hilt.navigation.compose.hiltViewModel
@Composable
fun TrainingScreen(
onSelectImages: () -> Unit,
modifier: Modifier = Modifier,
trainViewModel: TrainViewModel = hiltViewModel()
) {
var showInfoDialog by remember { mutableStateOf(false) }
Column(
modifier = modifier
.fillMaxSize()
.verticalScroll(rememberScrollState())
.padding(20.dp),
verticalArrangement = Arrangement.spacedBy(20.dp)
) {
// ✅ TIGHTENED Hero section
CompactHeroCard()
HowItWorksSection()
RequirementsCard()
Spacer(Modifier.weight(1f))
// Main CTA
Button(
onClick = { showInfoDialog = true },
modifier = Modifier.fillMaxWidth().height(60.dp),
colors = ButtonDefaults.buttonColors(containerColor = MaterialTheme.colorScheme.primary),
shape = RoundedCornerShape(16.dp)
) {
Icon(Icons.Default.PersonAdd, contentDescription = null, modifier = Modifier.size(24.dp))
Spacer(Modifier.width(12.dp))
Text("Start Training", style = MaterialTheme.typography.titleLarge, fontWeight = FontWeight.Bold)
}
Spacer(Modifier.height(8.dp))
}
// ✅ PersonInfo dialog BEFORE photo selection (CORRECT!)
if (showInfoDialog) {
BeautifulPersonInfoDialog(
onDismiss = { showInfoDialog = false },
onConfirm = { name, dob, relationship ->
showInfoDialog = false
trainViewModel.setPersonInfo(name, dob, relationship)
onSelectImages()
}
)
}
}
@Composable
private fun CompactHeroCard() {
Card(
modifier = Modifier.fillMaxWidth(),
colors = CardDefaults.cardColors(containerColor = MaterialTheme.colorScheme.primaryContainer),
shape = RoundedCornerShape(20.dp)
) {
Row(
modifier = Modifier
.fillMaxWidth()
.background(
Brush.horizontalGradient(
colors = listOf(
MaterialTheme.colorScheme.primaryContainer,
MaterialTheme.colorScheme.primaryContainer.copy(alpha = 0.7f)
)
)
)
.padding(20.dp),
horizontalArrangement = Arrangement.spacedBy(16.dp),
verticalAlignment = Alignment.CenterVertically
) {
// Compact icon
Surface(
shape = RoundedCornerShape(16.dp),
color = MaterialTheme.colorScheme.primary,
shadowElevation = 6.dp,
modifier = Modifier.size(56.dp)
) {
Box(contentAlignment = Alignment.Center) {
Icon(
Icons.Default.Face,
contentDescription = null,
modifier = Modifier.size(32.dp),
tint = MaterialTheme.colorScheme.onPrimary
)
}
}
// Text inline
Column(modifier = Modifier.weight(1f)) {
Text(
"Face Recognition",
style = MaterialTheme.typography.titleLarge,
fontWeight = FontWeight.Bold
)
Text(
"Train AI to find someone in your photos",
style = MaterialTheme.typography.bodyMedium,
color = MaterialTheme.colorScheme.onPrimaryContainer.copy(alpha = 0.8f)
)
}
}
}
}
@Composable
private fun HowItWorksSection() {
Column(verticalArrangement = Arrangement.spacedBy(12.dp)) {
Text("How It Works", style = MaterialTheme.typography.titleLarge, fontWeight = FontWeight.Bold)
StepCard(1, Icons.Default.Info, "Enter Person Details", "Name, birthday, and relationship")
StepCard(2, Icons.Default.PhotoLibrary, "Select Training Photos", "Choose 20-30 photos of the person")
StepCard(3, Icons.Default.SmartToy, "AI Training", "We'll create a recognition model")
StepCard(4, Icons.Default.AutoFixHigh, "Auto-Tag Photos", "Find this person across your library")
}
}
@Composable
private fun StepCard(number: Int, icon: ImageVector, title: String, description: String) {
Card(
modifier = Modifier.fillMaxWidth(),
colors = CardDefaults.cardColors(containerColor = MaterialTheme.colorScheme.surfaceVariant.copy(alpha = 0.5f)),
shape = RoundedCornerShape(16.dp)
) {
Row(
modifier = Modifier.padding(16.dp),
horizontalArrangement = Arrangement.spacedBy(16.dp),
verticalAlignment = Alignment.CenterVertically
) {
Surface(
modifier = Modifier.size(48.dp),
shape = RoundedCornerShape(12.dp),
color = MaterialTheme.colorScheme.primary
) {
Box(contentAlignment = Alignment.Center) {
Text("$number", style = MaterialTheme.typography.titleLarge, fontWeight = FontWeight.Bold, color = MaterialTheme.colorScheme.onPrimary)
}
}
Column(modifier = Modifier.weight(1f)) {
Row(horizontalArrangement = Arrangement.spacedBy(8.dp), verticalAlignment = Alignment.CenterVertically) {
Icon(icon, contentDescription = null, modifier = Modifier.size(20.dp), tint = MaterialTheme.colorScheme.primary)
Text(title, style = MaterialTheme.typography.titleMedium, fontWeight = FontWeight.SemiBold)
}
Text(description, style = MaterialTheme.typography.bodyMedium, color = MaterialTheme.colorScheme.onSurfaceVariant)
}
}
}
}
@Composable
private fun RequirementsCard() {
Card(
modifier = Modifier.fillMaxWidth(),
colors = CardDefaults.cardColors(containerColor = MaterialTheme.colorScheme.secondaryContainer.copy(alpha = 0.3f)),
shape = RoundedCornerShape(16.dp)
) {
Column(modifier = Modifier.padding(20.dp), verticalArrangement = Arrangement.spacedBy(12.dp)) {
Row(horizontalArrangement = Arrangement.spacedBy(8.dp), verticalAlignment = Alignment.CenterVertically) {
Icon(Icons.Default.CheckCircle, contentDescription = null, tint = MaterialTheme.colorScheme.primary, modifier = Modifier.size(24.dp))
Text("Best Results", style = MaterialTheme.typography.titleMedium, fontWeight = FontWeight.Bold)
}
RequirementItem(Icons.Default.PhotoCamera, "20-30 photos minimum")
RequirementItem(Icons.Default.Face, "Clear, well-lit face photos")
RequirementItem(Icons.Default.Diversity1, "Variety of angles & expressions")
RequirementItem(Icons.Default.HighQuality, "Good quality images")
}
}
}
@Composable
private fun RequirementItem(icon: ImageVector, text: String) {
Row(
horizontalArrangement = Arrangement.spacedBy(12.dp),
verticalAlignment = Alignment.CenterVertically,
modifier = Modifier.padding(vertical = 4.dp)
) {
Icon(icon, contentDescription = null, modifier = Modifier.size(20.dp), tint = MaterialTheme.colorScheme.onSecondaryContainer)
Text(text, style = MaterialTheme.typography.bodyMedium, color = MaterialTheme.colorScheme.onSecondaryContainer)
}
}

View File

@@ -0,0 +1,353 @@
package com.placeholder.sherpai2.ui.trainingprep
import androidx.compose.animation.AnimatedVisibility
import androidx.compose.foundation.BorderStroke
import androidx.compose.foundation.ExperimentalFoundationApi
import androidx.compose.foundation.background
import androidx.compose.foundation.combinedClickable
import androidx.compose.foundation.layout.*
import androidx.compose.foundation.lazy.grid.*
import androidx.compose.foundation.shape.CircleShape
import androidx.compose.foundation.shape.RoundedCornerShape
import androidx.compose.material.icons.Icons
import androidx.compose.material.icons.filled.*
import androidx.compose.material3.*
import androidx.compose.runtime.*
import androidx.compose.ui.Alignment
import androidx.compose.ui.Modifier
import androidx.compose.ui.draw.clip
import androidx.compose.ui.graphics.Color
import androidx.compose.ui.layout.ContentScale
import androidx.compose.ui.text.font.FontWeight
import androidx.compose.ui.unit.dp
import androidx.hilt.navigation.compose.hiltViewModel
import androidx.lifecycle.compose.collectAsStateWithLifecycle
import coil.compose.AsyncImage
import com.placeholder.sherpai2.data.local.entity.ImageEntity
/**
* TrainingPhotoSelectorScreen - Smart photo selector for face training
*
* SOLVES THE PROBLEM:
* - User has 10,000 photos total
* - Only ~500 have faces (hasFaces=true)
* - Shows ONLY photos with faces
* - Multi-select mode for quick selection
* - Face count badges on each photo
* - Minimum 15 photos enforced
*
* REUSES:
* - Existing ImageDao.getImagesWithFaces()
* - Existing face detection cache
* - Proven album grid layout
*/
@OptIn(ExperimentalMaterial3Api::class, ExperimentalFoundationApi::class)
@Composable
fun TrainingPhotoSelectorScreen(
onBack: () -> Unit,
onPhotosSelected: (List<android.net.Uri>) -> Unit,
viewModel: TrainingPhotoSelectorViewModel = hiltViewModel()
) {
val photos by viewModel.photosWithFaces.collectAsStateWithLifecycle()
val selectedPhotos by viewModel.selectedPhotos.collectAsStateWithLifecycle()
val isLoading by viewModel.isLoading.collectAsStateWithLifecycle()
Scaffold(
topBar = {
TopAppBar(
title = {
Column {
Text(
if (selectedPhotos.isEmpty()) {
"Select Training Photos"
} else {
"${selectedPhotos.size} selected"
},
style = MaterialTheme.typography.titleLarge,
fontWeight = FontWeight.Bold
)
Text(
"Showing ${photos.size} photos with faces",
style = MaterialTheme.typography.bodySmall,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
}
},
navigationIcon = {
IconButton(onClick = onBack) {
Icon(Icons.Default.ArrowBack, "Back")
}
},
actions = {
if (selectedPhotos.isNotEmpty()) {
TextButton(onClick = { viewModel.clearSelection() }) {
Text("Clear")
}
}
},
colors = TopAppBarDefaults.topAppBarColors(
containerColor = MaterialTheme.colorScheme.primaryContainer
)
)
},
bottomBar = {
AnimatedVisibility(visible = selectedPhotos.isNotEmpty()) {
SelectionBottomBar(
selectedCount = selectedPhotos.size,
onClear = { viewModel.clearSelection() },
onContinue = {
val uris = selectedPhotos.map { android.net.Uri.parse(it.imageUri) }
onPhotosSelected(uris)
}
)
}
}
) { paddingValues ->
Box(
modifier = Modifier
.fillMaxSize()
.padding(paddingValues)
) {
when {
isLoading -> {
Box(
modifier = Modifier.fillMaxSize(),
contentAlignment = Alignment.Center
) {
CircularProgressIndicator()
}
}
photos.isEmpty() -> {
EmptyState(onBack)
}
else -> {
PhotoGrid(
photos = photos,
selectedPhotos = selectedPhotos,
onPhotoClick = { photo -> viewModel.toggleSelection(photo) }
)
}
}
}
}
}
@Composable
private fun SelectionBottomBar(
selectedCount: Int,
onClear: () -> Unit,
onContinue: () -> Unit
) {
Surface(
modifier = Modifier.fillMaxWidth(),
color = MaterialTheme.colorScheme.primaryContainer,
shadowElevation = 8.dp
) {
Row(
modifier = Modifier
.fillMaxWidth()
.padding(16.dp),
horizontalArrangement = Arrangement.SpaceBetween,
verticalAlignment = Alignment.CenterVertically
) {
Column {
Text(
"$selectedCount photos selected",
style = MaterialTheme.typography.titleMedium,
fontWeight = FontWeight.Bold
)
Text(
when {
selectedCount < 15 -> "Need ${15 - selectedCount} more"
selectedCount < 20 -> "Good start!"
selectedCount < 30 -> "Great selection!"
else -> "Excellent coverage!"
},
style = MaterialTheme.typography.bodySmall,
color = when {
selectedCount < 15 -> MaterialTheme.colorScheme.error
else -> MaterialTheme.colorScheme.onPrimaryContainer.copy(alpha = 0.8f)
}
)
}
Row(horizontalArrangement = Arrangement.spacedBy(8.dp)) {
OutlinedButton(onClick = onClear) {
Text("Clear")
}
Button(
onClick = onContinue,
enabled = selectedCount >= 15
) {
Icon(
Icons.Default.Check,
contentDescription = null,
modifier = Modifier.size(20.dp)
)
Spacer(Modifier.width(8.dp))
Text("Continue")
}
}
}
}
}
@OptIn(ExperimentalFoundationApi::class)
@Composable
private fun PhotoGrid(
photos: List<ImageEntity>,
selectedPhotos: Set<ImageEntity>,
onPhotoClick: (ImageEntity) -> Unit
) {
LazyVerticalGrid(
columns = GridCells.Fixed(3),
contentPadding = PaddingValues(
start = 4.dp,
end = 4.dp,
bottom = 100.dp // Space for bottom bar
),
horizontalArrangement = Arrangement.spacedBy(4.dp),
verticalArrangement = Arrangement.spacedBy(4.dp)
) {
items(
items = photos,
key = { it.imageId }
) { photo ->
PhotoThumbnail(
photo = photo,
isSelected = photo in selectedPhotos,
onClick = { onPhotoClick(photo) }
)
}
}
}
@OptIn(ExperimentalFoundationApi::class)
@Composable
private fun PhotoThumbnail(
photo: ImageEntity,
isSelected: Boolean,
onClick: () -> Unit
) {
Card(
modifier = Modifier
.fillMaxWidth()
.aspectRatio(1f)
.combinedClickable(onClick = onClick),
shape = RoundedCornerShape(4.dp),
border = if (isSelected) {
BorderStroke(4.dp, MaterialTheme.colorScheme.primary)
} else null
) {
Box {
// Photo
AsyncImage(
model = photo.imageUri,
contentDescription = null,
modifier = Modifier.fillMaxSize(),
contentScale = ContentScale.Crop
)
// Face count badge (top-left)
if (photo.faceCount != null && photo.faceCount!! > 0) {
Surface(
modifier = Modifier
.align(Alignment.TopStart)
.padding(4.dp),
shape = CircleShape,
color = MaterialTheme.colorScheme.primaryContainer.copy(alpha = 0.95f)
) {
Row(
modifier = Modifier.padding(horizontal = 6.dp, vertical = 2.dp),
horizontalArrangement = Arrangement.spacedBy(2.dp),
verticalAlignment = Alignment.CenterVertically
) {
Icon(
Icons.Default.Face,
contentDescription = null,
modifier = Modifier.size(12.dp),
tint = MaterialTheme.colorScheme.onPrimaryContainer
)
Text(
"${photo.faceCount}",
style = MaterialTheme.typography.labelSmall,
color = MaterialTheme.colorScheme.onPrimaryContainer,
fontWeight = FontWeight.Bold
)
}
}
}
// Selection checkmark (top-right)
if (isSelected) {
Surface(
modifier = Modifier
.align(Alignment.TopEnd)
.padding(4.dp)
.size(28.dp),
shape = CircleShape,
color = MaterialTheme.colorScheme.primary,
shadowElevation = 4.dp
) {
Box(contentAlignment = Alignment.Center) {
Icon(
Icons.Default.CheckCircle,
contentDescription = "Selected",
modifier = Modifier.size(20.dp),
tint = MaterialTheme.colorScheme.onPrimary
)
}
}
}
// Dim overlay when selected
if (isSelected) {
Box(
modifier = Modifier
.fillMaxSize()
.background(Color.Black.copy(alpha = 0.2f))
)
}
}
}
}
@Composable
private fun EmptyState(onBack: () -> Unit) {
Box(
modifier = Modifier.fillMaxSize(),
contentAlignment = Alignment.Center
) {
Column(
horizontalAlignment = Alignment.CenterHorizontally,
verticalArrangement = Arrangement.spacedBy(16.dp),
modifier = Modifier.padding(32.dp)
) {
Icon(
Icons.Default.SearchOff,
contentDescription = null,
modifier = Modifier.size(72.dp),
tint = MaterialTheme.colorScheme.outline
)
Text(
"No Photos with Faces Found",
style = MaterialTheme.typography.titleLarge,
fontWeight = FontWeight.Bold
)
Text(
"Make sure the face detection cache has scanned your library",
style = MaterialTheme.typography.bodyMedium,
color = MaterialTheme.colorScheme.onSurfaceVariant
)
Button(onClick = onBack) {
Icon(Icons.Default.ArrowBack, null)
Spacer(Modifier.width(8.dp))
Text("Go Back")
}
}
}
}

View File

@@ -0,0 +1,116 @@
package com.placeholder.sherpai2.ui.trainingprep
import androidx.lifecycle.ViewModel
import androidx.lifecycle.viewModelScope
import com.placeholder.sherpai2.data.local.dao.ImageDao
import com.placeholder.sherpai2.data.local.entity.ImageEntity
import dagger.hilt.android.lifecycle.HiltViewModel
import kotlinx.coroutines.flow.MutableStateFlow
import kotlinx.coroutines.flow.StateFlow
import kotlinx.coroutines.flow.asStateFlow
import kotlinx.coroutines.launch
import javax.inject.Inject
/**
* TrainingPhotoSelectorViewModel - Smart photo selector for training
*
* KEY OPTIMIZATION:
* - Only loads images with hasFaces=true from database
* - Result: 10,000 photos → ~500 with faces
* - User can quickly select 20-30 good ones
* - Multi-select state management
*/
@HiltViewModel
class TrainingPhotoSelectorViewModel @Inject constructor(
private val imageDao: ImageDao
) : ViewModel() {
// Photos with faces (hasFaces=true)
private val _photosWithFaces = MutableStateFlow<List<ImageEntity>>(emptyList())
val photosWithFaces: StateFlow<List<ImageEntity>> = _photosWithFaces.asStateFlow()
// Selected photos (multi-select)
private val _selectedPhotos = MutableStateFlow<Set<ImageEntity>>(emptySet())
val selectedPhotos: StateFlow<Set<ImageEntity>> = _selectedPhotos.asStateFlow()
// Loading state
private val _isLoading = MutableStateFlow(true)
val isLoading: StateFlow<Boolean> = _isLoading.asStateFlow()
init {
loadPhotosWithFaces()
}
/**
* Load ONLY photos with hasFaces=true
*
* Uses indexed query: SELECT * FROM images WHERE hasFaces = 1
* Fast! (~10ms for 10k photos)
*/
private fun loadPhotosWithFaces() {
viewModelScope.launch {
try {
_isLoading.value = true
// ✅ CRITICAL: Only get images with faces!
val photos = imageDao.getImagesWithFaces()
// Sort by most faces first (better for training)
val sorted = photos.sortedByDescending { it.faceCount ?: 0 }
_photosWithFaces.value = sorted
} catch (e: Exception) {
// If face cache not populated, empty list
_photosWithFaces.value = emptyList()
} finally {
_isLoading.value = false
}
}
}
/**
* Toggle photo selection
*/
fun toggleSelection(photo: ImageEntity) {
val current = _selectedPhotos.value.toMutableSet()
if (photo in current) {
current.remove(photo)
} else {
current.add(photo)
}
_selectedPhotos.value = current
}
/**
* Clear all selections
*/
fun clearSelection() {
_selectedPhotos.value = emptySet()
}
/**
* Auto-select first N photos (quick start)
*/
fun autoSelect(count: Int = 25) {
val photos = _photosWithFaces.value.take(count)
_selectedPhotos.value = photos.toSet()
}
/**
* Select photos with single face only (best for training)
*/
fun selectSingleFacePhotos(count: Int = 25) {
val singleFacePhotos = _photosWithFaces.value
.filter { it.faceCount == 1 }
.take(count)
_selectedPhotos.value = singleFacePhotos.toSet()
}
/**
* Refresh data (call after face detection cache updates)
*/
fun refresh() {
loadPhotosWithFaces()
}
}

View File

@@ -0,0 +1,207 @@
package com.placeholder.sherpai2.ui.trainingprep
import android.content.Context
import android.graphics.Bitmap
import android.net.Uri
/**
* ENHANCED TrainingSanityChecker
*
* New features:
* - Progress callbacks
* - Exclude functionality
* - Faster processing
*/
class TrainingSanityChecker(private val context: Context) {
private val faceDetectionHelper = FaceDetectionHelper(context)
private val duplicateDetector = DuplicateImageDetector(context)
data class SanityCheckResult(
val isValid: Boolean,
val faceDetectionResults: List<FaceDetectionHelper.FaceDetectionResult>,
val duplicateCheckResult: DuplicateImageDetector.DuplicateCheckResult,
val validationErrors: List<ValidationError>,
val warnings: List<String>,
val validImagesWithFaces: List<ValidTrainingImage>,
val excludedImages: Set<Uri> = emptySet() // NEW: Track excluded images
)
data class ValidTrainingImage(
val uri: Uri,
val croppedFaceBitmap: Bitmap,
val faceCount: Int
)
sealed class ValidationError {
data class NoFaceDetected(val uris: List<Uri>) : ValidationError()
data class MultipleFacesDetected(val uri: Uri, val faceCount: Int) : ValidationError()
data class DuplicateImages(val groups: List<DuplicateImageDetector.DuplicateGroup>) : ValidationError()
data class InsufficientImages(val required: Int, val available: Int) : ValidationError()
data class ImageLoadError(val uri: Uri, val error: String) : ValidationError()
}
/**
* Perform comprehensive sanity checks with PROGRESS tracking
*/
suspend fun performSanityChecks(
imageUris: List<Uri>,
minImagesRequired: Int = 15,
allowMultipleFaces: Boolean = false,
duplicateSimilarityThreshold: Double = 0.95,
excludedImages: Set<Uri> = emptySet(), // NEW: Allow excluding images
onProgress: ((String, Int, Int) -> Unit)? = null // NEW: Progress callback
): SanityCheckResult {
val validationErrors = mutableListOf<ValidationError>()
val warnings = mutableListOf<String>()
// Filter out excluded images
val activeImages = imageUris.filter { it !in excludedImages }
// Check minimum image count (AFTER exclusions)
if (activeImages.size < minImagesRequired) {
validationErrors.add(
ValidationError.InsufficientImages(
required = minImagesRequired,
available = activeImages.size
)
)
}
// Step 1: Detect faces in all images (WITH PROGRESS)
onProgress?.invoke("Detecting faces...", 0, activeImages.size)
val faceDetectionResults = faceDetectionHelper.detectFacesInImages(
uris = activeImages,
onProgress = { current, total ->
onProgress?.invoke("Detecting faces...", current, total)
}
)
// Check for images without faces
val imagesWithoutFaces = faceDetectionResults.filter { !it.hasFace }
if (imagesWithoutFaces.isNotEmpty()) {
validationErrors.add(
ValidationError.NoFaceDetected(
uris = imagesWithoutFaces.map { it.uri }
)
)
}
// Check for images with errors
faceDetectionResults.filter { it.errorMessage != null }.forEach { result ->
validationErrors.add(
ValidationError.ImageLoadError(
uri = result.uri,
error = result.errorMessage ?: "Unknown error"
)
)
}
// Check for images with multiple faces
if (!allowMultipleFaces) {
faceDetectionResults.filter { it.faceCount > 1 }.forEach { result ->
validationErrors.add(
ValidationError.MultipleFacesDetected(
uri = result.uri,
faceCount = result.faceCount
)
)
}
} else {
faceDetectionResults.filter { it.faceCount > 1 }.forEach { result ->
warnings.add("Image ${result.uri.lastPathSegment} contains ${result.faceCount} faces. Using the largest detected face.")
}
}
// Step 2: Check for duplicate images
onProgress?.invoke("Checking for duplicates...", activeImages.size, activeImages.size)
val duplicateCheckResult = duplicateDetector.checkForDuplicates(
uris = activeImages,
similarityThreshold = duplicateSimilarityThreshold
)
if (duplicateCheckResult.hasDuplicates) {
validationErrors.add(
ValidationError.DuplicateImages(
groups = duplicateCheckResult.duplicateGroups
)
)
}
// Step 3: Create list of valid training images
val validImagesWithFaces = faceDetectionResults
.filter { it.hasFace && it.croppedFaceBitmap != null }
.filter { allowMultipleFaces || it.faceCount == 1 }
.map { result ->
ValidTrainingImage(
uri = result.uri,
croppedFaceBitmap = result.croppedFaceBitmap!!,
faceCount = result.faceCount
)
}
// Check if we have enough valid images after all checks
if (validImagesWithFaces.size < minImagesRequired) {
val existingError = validationErrors.find { it is ValidationError.InsufficientImages }
if (existingError == null) {
validationErrors.add(
ValidationError.InsufficientImages(
required = minImagesRequired,
available = validImagesWithFaces.size
)
)
}
}
val isValid = validationErrors.isEmpty() && validImagesWithFaces.size >= minImagesRequired
onProgress?.invoke("Analysis complete", activeImages.size, activeImages.size)
return SanityCheckResult(
isValid = isValid,
faceDetectionResults = faceDetectionResults,
duplicateCheckResult = duplicateCheckResult,
validationErrors = validationErrors,
warnings = warnings,
validImagesWithFaces = validImagesWithFaces,
excludedImages = excludedImages
)
}
/**
* Format validation errors into human-readable messages
*/
fun formatValidationErrors(errors: List<ValidationError>): List<String> {
return errors.map { error ->
when (error) {
is ValidationError.NoFaceDetected -> {
val count = error.uris.size
"No face detected in $count image(s)"
}
is ValidationError.MultipleFacesDetected -> {
"Multiple faces (${error.faceCount}) detected in: ${error.uri.lastPathSegment}"
}
is ValidationError.DuplicateImages -> {
val count = error.groups.size
"Found $count duplicate group(s)"
}
is ValidationError.InsufficientImages -> {
"Need ${error.required} images, have ${error.available}"
}
is ValidationError.ImageLoadError -> {
"Failed to load image: ${error.uri.lastPathSegment}"
}
}
}
}
/**
* Clean up resources
*/
fun cleanup() {
faceDetectionHelper.cleanup()
}
}

View File

@@ -0,0 +1,78 @@
package com.placeholder.sherpai2.ui.trainingprep
import android.app.Application
import android.net.Uri
import androidx.lifecycle.AndroidViewModel
import androidx.lifecycle.viewModelScope
import kotlinx.coroutines.flow.MutableStateFlow
import kotlinx.coroutines.flow.StateFlow
import kotlinx.coroutines.flow.asStateFlow
import kotlinx.coroutines.launch
/**
* ViewModel for managing training image sanity checks
*/
class TrainingSanityViewModel(application: Application) : AndroidViewModel(application) {
private val sanityChecker = TrainingSanityChecker(application)
private val _uiState = MutableStateFlow<TrainingSanityUiState>(TrainingSanityUiState.Idle)
val uiState: StateFlow<TrainingSanityUiState> = _uiState.asStateFlow()
sealed class TrainingSanityUiState {
object Idle : TrainingSanityUiState()
object Checking : TrainingSanityUiState()
data class Success(
val result: TrainingSanityChecker.SanityCheckResult
) : TrainingSanityUiState()
data class Error(val message: String) : TrainingSanityUiState()
}
/**
* Perform sanity checks on selected images
*/
fun checkImages(
imageUris: List<Uri>,
minImagesRequired: Int = 10,
allowMultipleFaces: Boolean = false,
duplicateSimilarityThreshold: Double = 0.95
) {
viewModelScope.launch {
try {
_uiState.value = TrainingSanityUiState.Checking
val result = sanityChecker.performSanityChecks(
imageUris = imageUris,
minImagesRequired = minImagesRequired,
allowMultipleFaces = allowMultipleFaces,
duplicateSimilarityThreshold = duplicateSimilarityThreshold
)
_uiState.value = TrainingSanityUiState.Success(result)
} catch (e: Exception) {
_uiState.value = TrainingSanityUiState.Error(
e.message ?: "An unknown error occurred during sanity checks"
)
}
}
}
/**
* Reset the UI state
*/
fun resetState() {
_uiState.value = TrainingSanityUiState.Idle
}
/**
* Get formatted error messages from validation result
*/
fun getFormattedErrors(result: TrainingSanityChecker.SanityCheckResult): List<String> {
return sanityChecker.formatValidationErrors(result.validationErrors)
}
override fun onCleared() {
super.onCleared()
sanityChecker.cleanup()
}
}

Some files were not shown because too many files have changed in this diff Show More