Skip to end of metadata
Go to start of metadata

You are viewing an old version of this page. View the current version.

Compare with Current View Page History

« Previous Version 7 Next »

SCHEDULE AT-A-GLANCE

DAY1. Monday, April 29
at ONE Summit 2024(On site only)


2 hour on-site discussion
  Agenda(Idea)
    Collaboration with other communities
    Outreach 
    Next Akraino release and 2024 activities
    AI at Edge



Day2. Monday, April 29 (APAC time zone friendly)


18:00 – 20:10 PDT (UTC-7)
21:00 – 23:10 EDT (UTC-4)
03:00 – 05:10 CEST (UTC+2) (Friday)
09:00 – 11:10 CST (UTC+8) (Friday)
Presentation Sessions


Day3. Wednesday, May 1 (APAC time zone friendly)
*Reserve day


18:00 – 20:10 PDT (UTC-7)
21:00 – 23:10 EDT (UTC-4)
03:00 – 05:10 CEST (UTC+2) (Friday)
09:00 – 11:10 CST (UTC+8) (Friday)
Presentation Sessions


Day2. 

Monday, April 29 (APAC time zone friendly)

Introduction to Akraino activities in 2023, Collaboration with other open communities

Zoom Link: TBD

Recording: TBD

Time(UTC-7)Topics
18:00-18:05

Welcome note
Yin Ding TSC Chair
Haruhisa Fukano TSC Co-Chair

18:10-18:40


18:40-19:10
19:10-19:40

19:40-20:10


Closing



Monday, May 1 (APAC time zone friendly)

Introduction to Akraino activities in 2023, Collaboration with other open communities

Zoom Link: TBD

Recording: TBD

Time(UTC-7)Topics
18:00-18:05

Welcome note
Yin Ding TSC Chair
Haruhisa Fukano TSC Co-Chair

18:10-18:40


18:40-19:10
19:10-19:40

19:40-20:10


Closing

Call for proposal

NoNameCompanye-mailPresentation titleAbstractionPreferred Time ZoneComments
1Jeff BrowerSignalogicjbrower at signalogic dot comSmall Language Model for Edge AI Applications

Abstract - Small Language Model for Device AI Applications

Device AI applications running on very small form-factor edge devices (for example pico ITX), and without a cloud connection, need to perform automatic speech recognition (ASR) under difficult conditions, including background noise, urgent or stressed voice input, and other talkers in the background. For robotics applications, background noise may also include servo motor and other mechanical noise. Under these conditions, ASRs such as Kaldi and Whisper tend to produce "sound alike" errors, for example:

  in the early days, a king rolled the stake
  
which contains two (2) sound-alike errors, and should be corrected to "in the early days a king ruled the state". Sound-alike errors are particularly problematic for robotics applications in which the robot OS requires precise commands, for example a stalled robotaxi must be instructed "move forward 20 feet, to the right 10 feet, raise the hood, and turn off the engine". A first responder may give these commands as "get off the road in that turn-out up ahead and shut it down" or similar. Any sound-alike errors in the first responder's commands make translation to machine commands problematic.

To address this problem, a Small Language Model (SLM) is needed to correct sound-alike errors, and capable of running in a very small form-factor and under 10W, for example using two (2) Atom CPU cores. The SLM must run every 1/2 second and with a backwards/forwards context of 3-4 words.

PDT
2






3






4






5






6






7






8






9






10



















  • No labels