iOS: alpha node app + setup-code onboarding (#11756)

This commit is contained in:
Mariano Belinky
2026-02-08 18:08:13 +01:00
parent 730f86dd5c
commit 6aedc54bd7
45 changed files with 6096 additions and 516 deletions

File diff suppressed because it is too large Load Diff

View File

@@ -7,6 +7,7 @@ struct TalkOrbOverlay: View {
var body: some View {
let seam = self.appModel.seamColor
let status = self.appModel.talkMode.statusText.trimmingCharacters(in: .whitespacesAndNewlines)
let mic = min(max(self.appModel.talkMode.micLevel, 0), 1)
VStack(spacing: 14) {
ZStack {
@@ -28,7 +29,7 @@ struct TalkOrbOverlay: View {
.fill(
RadialGradient(
colors: [
seam.opacity(0.95),
seam.opacity(0.75 + (0.20 * mic)),
seam.opacity(0.40),
Color.black.opacity(0.55),
],
@@ -36,6 +37,7 @@ struct TalkOrbOverlay: View {
startRadius: 1,
endRadius: 112))
.frame(width: 190, height: 190)
.scaleEffect(1.0 + (0.12 * mic))
.overlay(
Circle()
.stroke(seam.opacity(0.35), lineWidth: 1))
@@ -47,6 +49,13 @@ struct TalkOrbOverlay: View {
self.appModel.talkMode.userTappedOrb()
}
let agentName = self.appModel.activeAgentName.trimmingCharacters(in: .whitespacesAndNewlines)
if !agentName.isEmpty {
Text("Bot: \(agentName)")
.font(.system(.caption, design: .rounded).weight(.semibold))
.foregroundStyle(Color.white.opacity(0.70))
}
if !status.isEmpty, status != "Off" {
Text(status)
.font(.system(.footnote, design: .rounded).weight(.semibold))
@@ -59,6 +68,14 @@ struct TalkOrbOverlay: View {
.overlay(
Capsule().stroke(seam.opacity(0.22), lineWidth: 1)))
}
if self.appModel.talkMode.isListening {
Capsule()
.fill(seam.opacity(0.90))
.frame(width: max(18, 180 * mic), height: 6)
.animation(.easeOut(duration: 0.12), value: mic)
.accessibilityLabel("Microphone level")
}
}
.padding(28)
.onAppear {

View File

@@ -1,6 +1,7 @@
import AVFAudio
import Foundation
import Observation
import OpenClawKit
import Speech
import SwabbleKit
@@ -96,6 +97,7 @@ final class VoiceWakeManager: NSObject {
private var lastDispatched: String?
private var onCommand: (@Sendable (String) async -> Void)?
private var userDefaultsObserver: NSObjectProtocol?
private var suppressedByTalk: Bool = false
override init() {
super.init()
@@ -141,9 +143,28 @@ final class VoiceWakeManager: NSObject {
}
}
func setSuppressedByTalk(_ suppressed: Bool) {
self.suppressedByTalk = suppressed
if suppressed {
_ = self.suspendForExternalAudioCapture()
if self.isEnabled {
self.statusText = "Paused"
}
} else {
if self.isEnabled {
Task { await self.start() }
}
}
}
func start() async {
guard self.isEnabled else { return }
if self.isListening { return }
guard !self.suppressedByTalk else {
self.isListening = false
self.statusText = "Paused"
return
}
if ProcessInfo.processInfo.environment["SIMULATOR_DEVICE_NAME"] != nil ||
ProcessInfo.processInfo.environment["SIMULATOR_UDID"] != nil
@@ -159,14 +180,18 @@ final class VoiceWakeManager: NSObject {
let micOk = await Self.requestMicrophonePermission()
guard micOk else {
self.statusText = "Microphone permission denied"
self.statusText = Self.permissionMessage(
kind: "Microphone",
status: AVAudioSession.sharedInstance().recordPermission)
self.isListening = false
return
}
let speechOk = await Self.requestSpeechPermission()
guard speechOk else {
self.statusText = "Speech recognition permission denied"
self.statusText = Self.permissionMessage(
kind: "Speech recognition",
status: SFSpeechRecognizer.authorizationStatus())
self.isListening = false
return
}
@@ -364,20 +389,101 @@ final class VoiceWakeManager: NSObject {
}
private nonisolated static func requestMicrophonePermission() async -> Bool {
await withCheckedContinuation(isolation: nil) { cont in
AVAudioApplication.requestRecordPermission { ok in
cont.resume(returning: ok)
let session = AVAudioSession.sharedInstance()
switch session.recordPermission {
case .granted:
return true
case .denied:
return false
case .undetermined:
break
@unknown default:
return false
}
return await self.requestPermissionWithTimeout { completion in
AVAudioSession.sharedInstance().requestRecordPermission { ok in
completion(ok)
}
}
}
private nonisolated static func requestSpeechPermission() async -> Bool {
await withCheckedContinuation(isolation: nil) { cont in
SFSpeechRecognizer.requestAuthorization { status in
cont.resume(returning: status == .authorized)
let status = SFSpeechRecognizer.authorizationStatus()
switch status {
case .authorized:
return true
case .denied, .restricted:
return false
case .notDetermined:
break
@unknown default:
return false
}
return await self.requestPermissionWithTimeout { completion in
SFSpeechRecognizer.requestAuthorization { authStatus in
completion(authStatus == .authorized)
}
}
}
private nonisolated static func requestPermissionWithTimeout(
_ operation: @escaping @Sendable (@escaping (Bool) -> Void) -> Void) async -> Bool
{
do {
return try await AsyncTimeout.withTimeout(
seconds: 8,
onTimeout: { NSError(domain: "VoiceWake", code: 6, userInfo: [
NSLocalizedDescriptionKey: "permission request timed out",
]) },
operation: {
await withCheckedContinuation(isolation: nil) { cont in
Task { @MainActor in
operation { ok in
cont.resume(returning: ok)
}
}
}
})
} catch {
return false
}
}
private static func permissionMessage(
kind: String,
status: AVAudioSession.RecordPermission) -> String
{
switch status {
case .denied:
return "\(kind) permission denied"
case .undetermined:
return "\(kind) permission not granted"
case .granted:
return "\(kind) permission denied"
@unknown default:
return "\(kind) permission denied"
}
}
private static func permissionMessage(
kind: String,
status: SFSpeechRecognizerAuthorizationStatus) -> String
{
switch status {
case .denied:
return "\(kind) permission denied"
case .restricted:
return "\(kind) permission restricted"
case .notDetermined:
return "\(kind) permission not granted"
case .authorized:
return "\(kind) permission denied"
@unknown default:
return "\(kind) permission denied"
}
}
}
#if DEBUG