Skip Navigation
taters taters @lemmy.intai.tech
Posts 215
Comments 20
Artificial Intelligence - News | Events @lemmy.intai.tech taters @lemmy.intai.tech
slator.com How GPT-4 and Text-to-Speech Make Visual Content More Accessible

GPT-4 image processing and conversational AI combine with TTS to make visual content accessible for the blind and low-vision community.

How GPT-4 and Text-to-Speech Make Visual Content More Accessible
0
Artificial Intelligence - News | Events @lemmy.intai.tech taters @lemmy.intai.tech
www.engadget.com Google is testing its medical AI chatbot at the Mayo Clinic | Engadget

Google is already testing its Med-PaLM 2 AI chat technology at at the Mayo Clinic and other hospitals..

Google is testing its medical AI chatbot at the Mayo Clinic | Engadget
0

Curious Replay for Model-based Adaptation

Link: https://www.nature.com/articles/s41746-023-00873-0

Title: The Imperative for Regulatory Oversight of Large Language Models (or Generative AI) in Healthcare

Author(s): Bertalan Meskó & Eric J. Topol

Word count: 2,222

Estimated average read time: 10 minutes

Summary: This article emphasizes the need for regulatory oversight of large language models (LLMs) in healthcare. LLMs, such as GPT-4 and Bard, have the potential to revolutionize healthcare, but they also pose risks that must be addressed. The authors argue for differentiated regulation of LLMs in comparison to other AI-based medical technologies due to their unique characteristics and challenges.

The article discusses the scale, complexity, hardware requirements, broad applicability, real-time adaptation, societal impact, and data privacy concerns associated with LLMs. It highlights the need for a tailored regulatory approach that considers these factors. The authors also provide insights into the current regulatory landscape, particularly in the context of the United States' Food and Drug Administration (FDA), which has been adapting its framework to address AI and machine learning technologies in medical devices.

The authors propose practical recommendations for regulators, including the creation of a new regulatory category for LLMs, guidelines for deployment, consideration of future iterations with advanced capabilities, and focusing on regulating the companies developing LLMs rather than each individual model.

Evaluation for Applicability to Applications Development: This article provides valuable insights into the challenges and considerations regarding regulatory oversight of large language models in healthcare. While it specifically focuses on healthcare, the principles and recommendations discussed can be applicable to application development using large language models or generative AI systems in various domains.

Developers working on applications utilizing large language models should consider the potential risks and ethical concerns associated with these models. They should be aware of the need for regulatory compliance and the importance of transparency, fairness, data privacy, and accountability in their applications.

The proposed recommendations for regulators can also serve as a guide for developers, helping them shape their strategies for responsible and compliant development of applications using large language models. Understanding the regulatory landscape and actively addressing potential risks and challenges can lead to successful deployment and use of these models in different applications.

0

The imperative for regulatory oversight of large language models (or generative AI) in healthcare

www.nature.com The imperative for regulatory oversight of large language models (or generative AI) in healthcare - npj Digital Medicine

The rapid advancements in artificial intelligence (AI) have led to the development of sophisticated large language models (LLMs) such as GPT-4 and Bard. The potential implementation of LLMs in healthcare settings has already garnered considerable attention because of their diverse applications that ...

The imperative for regulatory oversight of large language models (or generative AI) in healthcare - npj Digital Medicine

Title: The Imperative for Regulatory Oversight of Large Language Models (or Generative AI) in Healthcare

Author(s): Bertalan Meskó & Eric J. Topol Word count: 2,222

Estimated average read time: 10 minutes

Summary: This article highlights the need for regulatory oversight of large language models (LLMs), such as GPT-4 and Bard, in healthcare settings. LLMs have the potential to transform healthcare by facilitating clinical documentation, summarizing research papers, and assisting with diagnoses and treatment plans. However, these models come with significant risks, including unreliable outputs, biased information, and privacy concerns.

The authors argue that LLMs should be regulated differently from other AI-based medical technologies due to their unique characteristics, including their scale, complexity, broad applicability, real-time adaptation, and potential societal impact. They emphasize the importance of addressing issues such as transparency, accountability, fairness, and data privacy in the regulatory framework.

The article also discusses the challenges of regulating LLMs, including the need for a new regulatory category, consideration of future iterations with advanced capabilities, and the integration of LLMs into already approved medical technologies.

The authors propose practical recommendations for regulators to bring this vision to reality, including creating a new regulatory category, providing guidance for deployment of LLMs, covering different types of interactions (text, sound, video), and focusing on companies developing LLMs rather than regulating each iteration individually.

Evaluation for Applicability to Applications Development: This article provides valuable insights into the regulatory challenges and considerations related to large language models in healthcare. While it primarily focuses on the medical field, the principles and recommendations discussed can be applicable to applications development using large language models or generative AI systems in various domains.

Developers working on applications that utilize large language models should be aware of the potential risks and ethical concerns associated with these models. They should also consider the need for regulatory compliance and the importance of transparency, fairness, data privacy, and accountability in their applications.

Additionally, developers may find the proposed practical recommendations for regulators helpful in shaping their own strategies for responsible and compliant development of applications using large language models. Understanding the regulatory landscape and being proactive in addressing potential risks and challenges can lead to the successful deployment and use of these models in various applications.

0
Artificial Intelligence - News | Events @lemmy.intai.tech taters @lemmy.intai.tech

George Hotz: Sam Altman won't tell you that GPT-4 has 220B parameters and is 16-way mixture model with 8 sets of weights

0
Artificial Intelligence - News | Events @lemmy.intai.tech taters @lemmy.intai.tech

Why the Senate may be accidentally giving carte blanche to potentially unethical AI

thehill.com Why the Senate may be accidentally giving carte blanche to unethical AI?

OpenAI’s CEO, Sam Altman, was at a Senate subcommittee hearing recently. This was a moment many were waiting for, and to their credit, the senators seemed to have done their homework and fostered a…

Why the Senate may be accidentally giving carte blanche to unethical AI?
0
Artificial Intelligence - News | Events @lemmy.intai.tech taters @lemmy.intai.tech

Andy Jassy dismisses Microsoft and Google A.I. ‘hype cycle’ and says Amazon is starting a ‘substance cycle’

https://archive.is/20230707210114/https://fortune.com/2023/07/07/andy-jassy-dismisses-microsoft-google-ai-hype-cycle-amazon-starting-substance-cycle/

0
Artificial Intelligence - News | Events @lemmy.intai.tech taters @lemmy.intai.tech

Grantham Warns AI Boom Won’t Prevent Market Bubble From Bursting

https://archive.is/20230701101545/https://www.wsj.com/articles/grantham-warns-ai-boom-wont-prevent-market-bubble-from-bursting-fd7f5a98

0
Artificial Intelligence - News | Events @lemmy.intai.tech taters @lemmy.intai.tech

Grammys boss Harvey Mason Junior clarifies AI policy and says 'music with AI-created elements is absolutely eligible for entry'

news.sky.com Grammys boss Harvey Mason Junior says music 'with AI-created elements' is 'absolutely eligible for entry'

As AI continues to make its presence felt in the music world, the Grammys have updated their rulebook. Now Recording Academy CEO Harvey Mason Junior has clarified rules around human vs AI-created work.

Grammys boss Harvey Mason Junior says music 'with AI-created elements' is 'absolutely eligible for entry'
0
Artificial Intelligence - News | Events @lemmy.intai.tech taters @lemmy.intai.tech

Google's updated privacy policy states it can use public data to train its AI models

www.engadget.com Google's updated privacy policy states it can use public data to train its AI models | Engadget

Google has updated its privacy policy to state that it can use publicly available data to help train its AI models..

Google's updated privacy policy states it can use public data to train its AI models | Engadget
0
New AI translates 5,000-year-old cuneiform tablets instantly
  • This is super cool. I've been thinking about using GPT to try to derive some meaning out of the Voynich manuscript, another one of a thousand projects I want to start

  • Hell You Talmbout - David Byrne's American Utopia

    0
    Artificial Intelligence - News | Events @lemmy.intai.tech taters @lemmy.intai.tech

    US AI Chip Controls Don’t Deter BlackRock From Going All In: The Week in AI

    https://archive.is/20230701122132/https://www.bloomberg.com/news/newsletters/2023-07-01/us-ai-chip-controls-blackrock-inflection-raises-1-3-billion-the-week-in-ai

    0
    Artificial Intelligence - News | Events @lemmy.intai.tech taters @lemmy.intai.tech

    Japan Dreams of AI, Overtaking Nvidia and Universal Basic Income

    https://archive.is/20230704074007/https://www.bloomberg.com/news/articles/2023-07-04/japan-dreams-of-ai-overtaking-nvidia-and-universal-basic-income

    0
    Artificial Intelligence - News | Events @lemmy.intai.tech taters @lemmy.intai.tech

    US to Curb China Access to Cloud Services Like Amazon, WSJ Says

    https://archive.is/20230704111538/https://www.bloomberg.com/news/articles/2023-07-04/us-to-curb-china-access-to-cloud-services-like-amazon-wsj-says

    0
    Instance Intros
  • Hey everyone! I'm Taters, one of the friendly admins here. Just a little about me: I'm a guy in my late 30's, a proud dog dad, and while I've lived in a half dozen cities across the country, I'm finally settled down out in the Southwest. In my past lives, I've been a touring comedian's personal sound engineer, managed a country music production studio in Nashville Tennessee, and even created a blockchain-based project to contribute to food assistance by supplying my state with tons (and I mean TONS) of potatoes (hence the name).

    I've been fascinated by AI science and theory ever since I was a kid and read Kurzweil's "Age of Spiritual Machines," which had a huge impact on me. Fast forward a couple of decades, and I've developed a passion for everything AI-related, across all sorts of vertical sectors. Lately, I've been focusing on prompt engineering with GPT-4 models, and I've discovered some incredible techniques thanks to the help of friends.

    I'm currently working on a few different projects in the areas of insurance, publishing and online learning. I'm excited to have an AI community to share all my news with.

    So, if you ever want to chat about AI, don't hesitate to reach out! I'm always up for discussing ethics, news, tools, or anything else AI-related. Looking forward to getting to know you all, and thanks for being here!

  • AI Made A Thing @lemmy.intai.tech taters @lemmy.intai.tech

    GPT + MJ + Photoshop, “invent new emotions” by @jphilipp

    1
    General Discussion @lemmy.intai.tech taters @lemmy.intai.tech

    Meta's "Threads", their twitter competitor app, shows up on the Apple app store

    0
    Artificial Intelligence - News | Events @lemmy.intai.tech taters @lemmy.intai.tech

    Artificial Intelligence Companies Hunt for San Francisco Offices

    https://archive.is/20230630231426/https://www.bloomberg.com/news/articles/2023-06-30/artificial-intelligence-firms-hunt-for-san-francisco-offices

    0
    Artificial Intelligence - News | Events @lemmy.intai.tech taters @lemmy.intai.tech

    Will AI Take My Job? What People Are Discussing Over The Holiday

    https://archive.is/20230703004526/https://www.bloomberg.com/news/articles/2023-07-03/will-ai-take-my-job-it-s-a-hot-topic-for-investors-this-summer#xj4y7vzkg

    1
    Artificial Intelligence - News | Events @lemmy.intai.tech taters @lemmy.intai.tech

    Bridgewater’s Greg Jensen Explains How the World’s Biggest Hedge Fund Is Investing in AI

    https://archive.is/20230703202352/https://www.bloomberg.com/news/articles/2023-07-03/bridgewater-s-greg-jensen-explains-how-the-world-s-biggest-hedge-fund-is-investing-in-ai#selection-3781.0-3781.88

    0
    Natural Language Programming | Prompting (chatGPT) @lemmy.intai.tech taters @lemmy.intai.tech

    Best Prompt Generating/Refinement Prompt

    Act as a prompt generator for ChatGPT. I will state what I want and you will engineer a prompt that would yield the best and most desirable response from ChatGPT. Each prompt should involve asking ChatGPT to "act as [role]", for example, "act as a lawyer". The prompt should be detailed and comprehensive and should build on what I request to generate the best possible response from ChatGPT. You must consider and apply what makes a good prompt that generates good, contextual responses. Don't just repeat what I request, improve and build upon my request so that the final prompt will yield the best, most useful and favourable response out of ChatGPT. Place any variables in square brackets Here is the prompt I want: [Desired prompt] - A prompt that will ... Ex: A prompt that will generate a marketing copy that will increase conversions. Start by asking the user what they want the prompt to be about.

    0
    Fediverse/Lemmy Resources
  • Permanently Deleted

  • Maxine Waters: 'We don't know the dangers of AI'
  • shouldn't be able to make laws about things you can't explain

  • Ameca: The Future Face of AI Robots, Drawing a cat
  • And then the snarky comment at the end lol

  • How the FUCK is this not the Lemmy Logo?
  • seems like a no brainer to me

  • AT&T unveils "Ask AT&T", A New ChatGPT based AI Tool That Will Help Employees Do Their Job More Effectively
  • Absolutely. Creating more efficient Customer Service Representatives will lead to needing less workers. As a publicly traded company the burdens on ATT to make as much money as possible for their shareholders, so layoffs are bound to happen. Company I work for will be doing the same, I'm sure of it.

  • 20% of Professionals think ChatGPT Should be Banned in the Workplace
  • There 's a decent amount of people I work with who have that irrational fear of AI because they don't understand it. It's unfortunate.

  • Judging LLM-as-a-judge with MT-Bench and Chatbot Arena
  • Title: Judging LLM-as-a-judge with MT-Bench and Chatbot Arena

    **Authors: Lianmin Zheng et al.

    Word Count: Approximately 6,400 words

    Estimated Read Time: 22-24 minutes

    Summary:

    The paper proposes using strong LLMs as judges to evaluate LLM-based chat assistants in a scalable way. The authors examine the use of LLM-as-a-judge by looking at position bias, verbosity bias, and limited reasoning ability. They evaluate LLM judges using two benchmarks: MT-Bench, a multi-turn question set, and Chatbot Arena, a crowdsourced platform.

    The results show that GPT-4 achieves over 80% agreement with human preferences, matching the level of human-human agreement. This suggests that LLM-as-a-judge is a promising alternative to costly human evaluations.

    The authors explore variations of LLM-as-a-judge: pairwise comparison, single answer grading, and chain-of-thought/reference-guided judging. They also examine finetuning a Vicuna base model as a judge.

    The authors release MT-Bench with 80 questions and Chatbot Arena data comprising 30K conversations. They argue for a hybrid evaluation framework combining standardized benchmarks and LLM-as-a-judge evaluations.

    Evaluation of variants with MMLU, TruthfulQA, and MT-Bench show that benchmarks capture complementary aspects, indicating the need for a comprehensive evaluation approach.

    In summary, the paper provides empirical evidence that LLMs can serve as scalable proxies for human preferences in chatbot evaluation. However, further work is needed to mitigate biases and improve LLM judging models.

    Potential Use: LLM-as-a-judge can enable fast, automated assessments of LLMs' helpfulness, relevance and instruction-following ability in human-aligned dialogue systems. The proposed benchmarks and finetuning methods can be used to improve existing dialogue models.

  • CREPE - Can Vision-Language Foundation Models Reason Compositionally?
  • Title: CREPE: Can Vision-Language Foundation Models Reason Compositionally?

    Authors: Zixian Ma, Jerry Hong, Mustafa Omer Gul, Mona Gandhi, Irena Gao, Ranjay Krishna

    Word Count: Approximately 7,600 words

    Estimated Read Time: 25-30 minutes

    Summary:

    The paper introduces CREPE, a benchmark for evaluating the compositional reasoning abilities of vision-language foundation models. Compositionality refers to the ability to understand and generate complex visual scenes or statements by combining simpler parts. The benchmark covers two important aspects of compositionality: systematicity and productivity.

    Systematicity evaluates a model's ability to systematically recombine known visual concepts in unseen combinations. CREPE includes three systematicity splits based on whether models have seen all concepts ("Seen Compounds"), only the individual concepts ("Unseen Compounds"), or neither ("Unseen Atoms"). CREPE finds that most models' performance decreases when evaluating on unseen combinations of concepts, especially for models trained on the larger LAION-400M dataset.

    Productivity evaluates a model's ability to comprehend visual concepts of increasing compositional complexity. CREPE includes captions ranging from 4 to 12 visual concepts. It finds that most models struggle on captions with higher compositional complexity, with retrieval performance nearing random chance.

    Overall, CREPE finds that vision-language foundation models trained with contrastive loss struggle at compositional reasoning, even when trained on large datasets and using large model architectures. CREPE aims to provide a comprehensive benchmark to track the emergence of compositionality as vision-language models improve.

    CREPE provides large-scale datasets with ground truth image-caption pairs and hard negative captions to evaluate both systematicity and productivity. Hard negatives differ from ground truth captions in minimal ways to isolate model failure modes.

    Researchers can use CREPE to identify gaps in current foundation models relating to compositionality. Improving compositionality could make models more controllable and robust. CREPE's hard negative generation approach could also be used to improve the training of compositional models.

    CREPE relies on a scene graph representation to define compositional language. The generated hard negatives are noisy, especially swapping and negation foils. Evaluating productivity also relies on generated captions.

  • Human-Like Intuitive Behavior and Reasoning Biases Emerged in Language Models—and Disappeared in GPT-4
  • Title: Machine intuition: Uncovering human-like intuitive decision-making in GPT-3

    Authors: Thilo Hagendorff, Sarah Fabi , and Michal Kosinski

    Word Count: Approximately 10,200 words

    Estimated Read Time: 35-40 minutes

    Summary:

    The paper investigates whether large language models (LLMs) like GPT-3 exhibit behaviors similar to human intuition and cognitive biases. The authors probe various LLMs with the Cognitive Reflection Test (CRT) and semantic illusions, which were originally designed to study intuitive decision-making in humans.

    The results show that early LLMs lack the mathematical abilities and knowledge to perform these tasks. However, as LLMs become more complex, they begin to show human-like intuitive behavior and make the same errors as humans. GPT-3 in particular exhibits a strong inclination for intuitive responses on the CRT and semantic illusions, responding correctly in only around 10% of cases.

    However, newer LLMs like ChatGPT and GPT-4 overcome these intuitive errors, responding correctly in around 80% and 97% of cases respectively. The authors attribute this to increases in ChatGPT and GPT-4's reasoning capabilities.

    The authors explore methods to reduce intuitive behavior in GPT-3, such as providing multiple choice options, eliciting deliberate reasoning, and providing training examples. These methods are effective, bringing GPT-3's performance close to ChatGPT's.

    The findings suggest that LLMs can develop probability distributions over language that mimic human intuition, even though they lack cognitive mechanisms. The authors argue that investigating LLMs with methods from psychology has the potential to reveal otherwise unknown behavior.

    In summary, the paper demonstrates that LLMs gradually develop the ability to make human-like intuitive decisions and errors. However, the newest LLMs seem to overcome these tendencies, suggesting major improvements in their reasoning capabilities. The findings highlight the value of using methods from psychology to study the abilities and behaviors of LLMs.

    The findings could inform the development of LLMs that are designed to avoid intuitive errors and be more robust reasoners. The methods used to study human-like behavior in LLMs could also be applied to new models as they are developed. The results also highlight the need for careful scrutiny of LLMs before deploying them in real-world applications.

  • Model Sketching - Centering Concepts in Early-Stage Machine Learning Model Design
  • Title: Model Sketching: Centering Concepts in Early-Stage Machine Learning Model Design

    Authors: Michelle S. Lam, Zixian Ma, Anne Li, Izequiel Freitas, Dakuo Wang, James A. Landay, and Michael S. Bernstein

    Word Count: Approximately 24,000 words

    Estimated Read Time: 80-90 minutes

    Source Code/Repositories: ModelSketchBook API code - https://github.com/StanfordHCI/ModelSketchBook

    Summary:

    The paper introduces the concept of model sketching for early-stage machine learning model design. Model sketching allows ML practitioners to prototype model behavior through lightweight sketches that focus on high-level concepts relevant to a decision-making task.

    The key ideas are:

    • Concepts: Human-understandable factors that a model reasons over, like "profanity" or "sarcasm". Concepts serve as functional building blocks of model logic.

    • Zero-shot concept instantiation: Models like GPT-3 and CLIP are leveraged to flexibly and rapidly instantiate concepts without diverting user attention.

    • Sketch models: Composite models that aggregate concept scores, allowing ML practitioners to explore different combinations of concepts.

    The authors implemented model sketching in ModelSketchBook, an open-source Python API. An evaluation with 17 ML practitioners found that model sketching shifted their focus from technical implementation to higher-level concepts. This cognitive shift helped participants explore a broader model design space and identify gaps in data, labels, and problem formulation.

    The paper argues that model sketching can help ML practitioners move beyond reactive fixes and engage in more proactive model design exploration from the start.

    Overall, model sketching allows ML practitioners to rapidly prototype different model design possibilities through lightweight sketches centered on human-understandable concepts. This represents a shift away from technical tunneling towards higher-level conceptual thinking during early model design stages.

  • WebGLM - Towards An Efficient Web-Enhanced Question Answering System with Human Preferences
  • Paper Title: GPT Understands, Too

    Authors: Xiao Liu, Yanan Zheng, Zhengxiao Du, Ming Ding, Yujie Qian, Zhilin Yang, Jie Tang

    Word Count: 13,000 words

    Estimated Read Time: 40-45 minutes

    Source Code/Repositories:

    Summary:

    The paper studies GPT-3's capabilities beyond language generation, finding that GPT-3 has the ability to understand knowledge and commonsense reasoning despite its generative pre-training objective.

    The key findings are:

    1. GPT-3 can perform knowledge verification tasks with high accuracy, detecting factual errors in 95.5% of cases.

    2. GPT-3 can infer correct results from premises in 88.6% of cases on a causal reasoning task.

    3. GPT-3 demonstrates systematicity in its reasoning, generalizing causal rules to novel contexts.

    4. GPT-3 shows dose-response behavior, with performance increasing as the number of evidence sentences increases.

    5. GPT-3's performance is relatively robust to the number of facts and details in a given context.

    The authors argue that GPT-3's knowledge and reasoning capabilities emerge from its autoregressive pre-training objective, which implicitly forces the model to capture dependencies between words to predict the next token.

    In summary, the paper provides compelling evidence that large language models like GPT-3 have acquired substantial capabilities beyond text generation, posing new opportunities and challenges for deploying and scrutinizing these powerful systems.

    The findings suggest that generative pre-training objectives can implicitly teach language models to perform tasks like knowledge verification and commonsense reasoning, without being optimized for those specific goals. This suggests large language models may become a promising foundation for building AI applications with more comprehensive capabilities.

  • Self-Supervised Learning from Images with a Joint-Embedding Predictive Architecture
  • Title: Self-Supervised Learning from Images with a Joint-Embedding Predictive Architecture

    Authors: Mahmoud Assran, Quentin Duval, Ishan Misra, Piotr Bojanowski, Pascal Vincent, Michael Rabbat, Yann LeCun, Nicolas Ballas

    Word Count: Approximately 10,200 words

    Estimated Read Time: 35-40 minutes

    Source Code/Repositories: Not mentioned

    Links: Not applicable

    Summary: This paper proposes a joint-embedding predictive architecture called I-JEPA for self-supervised learning of visual representations from images. Traditional self-supervised learning approaches involve either view-invariance methods that require hand-crafted data augmentations or generative methods that require pixel-level reconstruction. I-JEPA predicts missing information in representation space instead of pixel space, which allows it to learn more semantic features.

    A key design choice is the multi-block masking strategy that samples sufficiently large target blocks and an informative context block. Experiments show that I-JEPA learns strong representations without data augmentations and outperforms pixel-reconstruction methods. It also demonstrates better performance on low-level tasks compared to view-invariance methods. I-JEPA also has better scalability due to its efficiency, requiring less computation compared to previous methods.

    Applicability: The I-JEPA approach could be applicable for developing self-supervised vision models using large language models or GANs. Predicting in representation space rather than pixel space allows the model to learn more semantic features, which could be beneficial for language models. The scalability and efficiency of I-JEPA is also promising for scaling to large models. Key ideas like the multi-block masking strategy and importance of semantic target blocks could be useful design principles. However, directly applying I-JEPA to language models or GANs would likely require significant adaptations. The paper mainly focuses on proving the concept of abstract representation-space prediction for self-supervised learning in vision.

    Overall, the key ideas and findings regarding abstract prediction targets, masking strategies, and scalability could inspire self-supervised methods for developing vision components of multimodal models, powered by large language models or GANs. But directly applying the I-JEPA approach would require addressing challenges specific to those modalities and applications.