@chikim@mastodon.social avatar

chikim

@chikim@mastodon.social

Love music, technology, accessibility! Faculty at Berklee College of Music ๐Ÿ‘จ๐Ÿปโ€๐Ÿ’ป๐ŸŽน๐Ÿ•โ€๐Ÿฆบ

This profile is from a federated server and may be incomplete. Browse more on the original instance.

chikim, to random
@chikim@mastodon.social avatar

BeMyEyes Privacy Policy 1/2: We record and store video streams and other images to enforce our Terms of Service, to promote and preserve safety, and to improve our Services and create new Services. We may provide recorded video streams or images to other organizations that are performing research or working to develop products and services that may assist blind and low-vision people or other members of the general public.

chikim,
@chikim@mastodon.social avatar

BeMyEyes Privacy Policy 2/2: If you use Be My AI, the images you submit will be processed by our third party artificial intelligence provider. If your video or image contains personal information - for example you say your name on the video or show mail with your home address - that information will be included in videos or images that we store and use as described above.

chikim,
@chikim@mastodon.social avatar

Do not feed images of your online meetings to BeMyAI unless you have consent from everyone involved to use their faces and names for AI training. https://www.bemyeyes.com/privacy

chikim,
@chikim@mastodon.social avatar

@JonathanMosen Sorry for specifically tagging you, but do you have any thought on the BeMyEyes privacy policy re images and videos you submit to the platform? It's too long to include in one post, but I pasted the relevant quotes in this thread. Should blind folks just say no such thing as free lunch and move on?

chikim,
@chikim@mastodon.social avatar

@twynn @JonathanMosen For paid customer, OpenAI has feature to opt out from getting your data used for training. Also if you use API, your data "do not become part of the training data unless you explicitly opt in." Your data is deleted within 30 days unless required for legal reasons, and is only accessible by authorized OpenAI employees, as well as specialized third-party contractors (that are subject to confidentiality and security obligations). https://www.maginative.com/article/openai-clarifies-its-data-privacy-practices-for-api-users/

chikim,
@chikim@mastodon.social avatar

@twynn @JonathanMosen Actually, if you have a free OpenAI account, you can turn off Improve the model for everyone in Settings > Data controls on this webpage. How much can we trust them? It's different story. lol https://chatgpt.com/#settings/DataControls

chikim, to random
@chikim@mastodon.social avatar

Finally release VOCR 2.0.0. So many new features since 1.0! You can download and checkout the demo here. https://chigkim.github.io/VOCR/

chikim, to llm
@chikim@mastodon.social avatar
chikim, to random
@chikim@mastodon.social avatar

"OpenAI has paused a voice mode option for ChatGPT-4o, Sky, after backlash accusing the AI company of intentionally ripping off Scarlett Johansson's critically acclaimed voice-acting performance in the 2013 sci-fi film Her." I don't think Sky sounds like Scarlett Johansson, but whatever. https://arstechnica.com/tech-policy/2024/05/openai-pauses-chatgpt-4o-voice-that-fans-said-ripped-off-scarlett-johansson/

vick21, to accessibility
@vick21@mastodon.social avatar

How NVDA & OSARA are empowering blind people globally - Audio described Version: https://youtube.com/watch?v=N-y3yomLLSk&si=xiibf5ZxJzrlDnES

chikim,
@chikim@mastodon.social avatar

@vick21 I understand this is not a good way to measure by any mean, but regardless let's do some math! Globally there are 43M blind people , According to WHO. There are over 250k NVDA users in 175 countries, According to NVDA creators. Let's just say 300k. According to screen reader survey , NVDA counts for 37.7%. That means only 1.85% of blind people have access to screen reader. 0.3M/0.377%/43M*100 That's very sad! :( Let me know if I epically failed this math. lol

chikim,
@chikim@mastodon.social avatar

@miki @pixelate @vick21 Good point! Globally 82% of blind people are over 50. If you adjust the stat to look only under 50 years old, it's 11.33%. I think also screen reader survey includes screen readers on mobile devices, so it's not just computer.

chikim, to random
@chikim@mastodon.social avatar

I created samples for all 58 voices for xtts-v2. Hopefully it makes it easier for someone to choose a speaker. #coqui #tts https://we.tl/t-9vWd1gO3EN

chikim, to apple
@chikim@mastodon.social avatar

Am I the only one who hates the giant escape key on newer Macbook Pro? It pushes all the function keys to the right, and it really throws off my muscle memory! Also Logitech thinks it's a great idea on MX Keys S. I just got older Logitech MX Mini for this reason. :(

chikim, to ai
@chikim@mastodon.social avatar

lol Cheery! The best name for ChatGPT 4O voice! Jeff Jarvis on TWIT podcast suggested. I agree, she's so over the top!

blindbargains, to random
@blindbargains@mastodon.social avatar

Be My Eyes Accessibility with GPT-4o https://www.youtube.com/watch?v=KwNUJ69RbwY

chikim,
@chikim@mastodon.social avatar

@FreakyFwoof @blindbargains Is gpt-4O on BeMyAI available now? Only for Beta users?

chikim, to llm
@chikim@mastodon.social avatar

Interesting, the Chat GPT desktop app for the Mac will be slowly rolling out to Plus subscribers starting today, but OpenAI "plans to release a Windows version of the desktop app later this year." Maybe the rumor, that Apple is cosing a deal with OpenAi for ChatGPT, is True... https://www.macrumors.com/2024/05/13/chatgpt-mac-app/

chikim, to llm
@chikim@mastodon.social avatar

If you missed it, check out the new GPT4O demo. #LLM #ML #AI #OpenAI https://www.youtube.com/watch?v=DQacCB9tDaw

chikim,
@chikim@mastodon.social avatar

@bryansmart lol I'm the one who posted about Google IO, and you replied to me. haha

chikim,
@chikim@mastodon.social avatar

@bryansmart FAlso funny. Google IO event is tomorrow. OpenAI intentionally revealed their model today. lol

chikim, to llm
@chikim@mastodon.social avatar

GPT4O combines audio, image, text. It can analyze actual audio. Also you can interrupt voice. It can pick up emotion from audio. You can also ask different speech with diffetrent style including singing! It can see the image in real time and chat on voice. For example solving equation in real time as you write on paper. This is amazing!

chikim, to accessibility
@chikim@mastodon.social avatar

I'm late to the party, but I found out I'm with the majority! lol According to the WebAIM Screen Reader User Survey , 68.2% (779 out of 1142) "respondents indicate that individuals should not describe what they look like during a virtual meeting or webinar" for blind and visually impaired participants. https://www.webaxe.org/webaim-screen-reader-user-survey-10/

chikim, to ai
@chikim@mastodon.social avatar

ElevenLabs joined the AI music generation. not available for public, but here's a demo clip. It's getting better and better! #AI #ML https://www.youtube.com/watch?v=m9DrkOrr3QM

chikim, to random
@chikim@mastodon.social avatar

Well Logic Pro on Ipad now has built in stem separation like Demucs. lol

macrumors, to random
@macrumors@mastodon.social avatar
chikim,
@chikim@mastodon.social avatar

@macrumors Hi, do you guys have MacRumorsLive on Mastodon?

bryansmart, to random

@chikim I've really been enjoying VOLlama. Nice work! Would be nice to be able to switch between OpenAI and local models without going in to API prefs. More accelerator keys for menu options would be good, too. Could maybe a blank line be inserted in the log between each entry? Last, can you trap key-down on the Control key to stop the system voice? I know it's a hobby project, so no idea how much time you have for any of that, but just throwing them out there.

chikim,
@chikim@mastodon.social avatar

@bryansmart Embedding model is only used during indexing. The quality of the answer depends on the model you are chatting with. Because it will read the chunks in text and give you the answer. It really depends on if the LlamaIndex was able to retrieve the relevant chunks or not. You can increase number of chunks and chunk length, but you might end up feeding chunks that are not related to your question. Also there's threshold you can change to filter out below certain simularity score.

  • All
  • Subscribed
  • Moderated
  • Favorites
  • โ€ข
  • JUstTest
  • kavyap
  • DreamBathrooms
  • thenastyranch
  • magazineikmin
  • InstantRegret
  • Durango
  • Youngstown
  • everett
  • slotface
  • rosin
  • cubers
  • mdbf
  • ngwrru68w68
  • anitta
  • GTA5RPClips
  • cisconetworking
  • osvaldo12
  • ethstaker
  • Leos
  • khanakhh
  • normalnudes
  • tester
  • modclub
  • tacticalgear
  • megavids
  • provamag3
  • lostlight
  • All magazines