The iOS app which connects to the Listener desk accessory on an Apple IIgs for doing speech dictation.
Go to file
2022-03-23 22:41:25 -04:00
ListenerGS Change the write side into a series of short running operations rather than a single operation looping while listening. 2022-03-23 22:41:25 -04:00
ListenerGS.xcodeproj Add an initial help screen that is visible on iPad and macOS. 2022-03-16 20:37:18 -04:00
ListenerGSTests Change the write side into a series of short running operations rather than a single operation looping while listening. 2022-03-23 22:41:25 -04:00
ListenerGSUITests Major rework of the network code, splitting out the speech recognition aspects and creating separate threads for read and write. This improves handling of network connection closure by the other end and other network connectivity errors. Add some unit tests for the connection code. 2022-03-16 18:13:56 -04:00
LICENSE
README.md Add an initial help screen that is visible on iPad and macOS. 2022-03-16 20:37:18 -04:00

ListenerGS

This is an iOS app that connects to an Apple IIgs over a network and streams text to it from voice dicatation. It communicates to the Listen NDA which must be running on your network capable Apple IIgs. See that other project for more details about how to use this app.

Obtaining a Copy

I am getting close to making this available as a release that anyone can obtain. It is currently in beta test and the iOS and macOS app have passed Apple's beta testing review. That doesn't mean that the app will pass full review but I am hopeful. So, perhaps soon I will be able to provide a link for where to download this for yourself.

Some Technical Details

It relies on the Speech Framework built into recent versions of iOS.

This app uses:

  • SwiftSocket v2.1.0 for opening the TCP connection to the Apple IIgs.
  • BinUtils for packing/unpacking structures on the TCP connection.
  • RichText v1.7.0 for displaying a nice startup screen on iPad and macOS.

Warning

This is my first from scratch application in Swift and in SwiftUI and it shows. I know this code is crufty and bad, even as an inexperienced Swift coder. It was created as part of the KansasFest 2021 HackFest contest so I only had so much time for code cleanup. Since then, I have done a bunch of cleanup and things are better. But I am still not happy with how the code handles the network reads and writes. A producer/consumer approach where the speech recognizer produces text and the network code consumes that text would be better. But Apple just changed the multithreading approach in Swift this past year and I am not sure the best way to structure that. Be patient as this old dog tries to learn some new tricks.