home.social

#aied — Public Fediverse posts

Live and recent posts from across the Fediverse tagged #aied, aggregated by home.social.

  1. Self-reported measures (surveys) are often not correlated or even negatively correlated w/more objective measures (such as observations, scenario/performance assessments). Examples:
    * Teacher AI literacy arxiv.org/abs/2601.06101
    * Applying professional development to the classroom academic.oup.com/bioscience/ar
    * AI cognitive offloading goedel.io/p/the-machine-that-s
    * Student learning from teaching pnas.org/doi/10.1073/pnas.1821
    * And grades link.springer.com/article/10.1
    * TPACK osf.io/preprints/psyarxiv/bhqx
    #EdDev #AIEd

  2. Self-reported measures (surveys) are often not correlated or even negatively correlated w/more objective measures (such as observations, scenario/performance assessments). Examples:
    * Teacher AI literacy arxiv.org/abs/2601.06101
    * Applying professional development to the classroom academic.oup.com/bioscience/ar
    * AI cognitive offloading goedel.io/p/the-machine-that-s
    * Student learning from teaching pnas.org/doi/10.1073/pnas.1821
    * And grades link.springer.com/article/10.1
    * TPACK osf.io/preprints/psyarxiv/bhqx
    #EdDev #AIEd

  3. Self-reported measures (surveys) are often not correlated or even negatively correlated w/more objective measures (such as observations, scenario/performance assessments). Examples:
    * Teacher AI literacy arxiv.org/abs/2601.06101
    * Applying professional development to the classroom academic.oup.com/bioscience/ar
    * AI cognitive offloading goedel.io/p/the-machine-that-s
    * Student learning from teaching pnas.org/doi/10.1073/pnas.1821
    * And grades link.springer.com/article/10.1
    * TPACK osf.io/preprints/psyarxiv/bhqx
    #EdDev #AIEd

  4. Self-reported measures (surveys) are often not correlated or even negatively correlated w/more objective measures (such as observations, scenario/performance assessments). Examples:
    * Teacher AI literacy arxiv.org/abs/2601.06101
    * Applying professional development to the classroom academic.oup.com/bioscience/ar
    * AI cognitive offloading goedel.io/p/the-machine-that-s
    * Student learning from teaching pnas.org/doi/10.1073/pnas.1821
    * And grades link.springer.com/article/10.1
    * TPACK osf.io/preprints/psyarxiv/bhqx
    #EdDev #AIEd

  5. Self-reported measures (surveys) are often not correlated or even negatively correlated w/more objective measures (such as observations, scenario/performance assessments). Examples:
    * Teacher AI literacy arxiv.org/abs/2601.06101
    * Applying professional development to the classroom academic.oup.com/bioscience/ar
    * AI cognitive offloading goedel.io/p/the-machine-that-s
    * Student learning from teaching pnas.org/doi/10.1073/pnas.1821
    * And grades link.springer.com/article/10.1
    * TPACK osf.io/preprints/psyarxiv/bhqx
    #EdDev #AIEd

  6. Designing a mobile chatbot-based learning journaling system for intrinsic motivation and engagement link.springer.com/article/10.1... #AIEd #Education #EdTech

    Designing a mobile chatbot-bas...

  7. Designing a mobile chatbot-based learning journaling system for intrinsic motivation and engagement
    link.springer.com/article/10.1
    #AIEd #Education #EdTech

  8. Designing a mobile chatbot-based learning journaling system for intrinsic motivation and engagement
    link.springer.com/article/10.1
    #AIEd #Education #EdTech

  9. Designing a mobile chatbot-based learning journaling system for intrinsic motivation and engagement
    link.springer.com/article/10.1
    #AIEd #Education #EdTech

  10. Designing a mobile chatbot-based learning journaling system for intrinsic motivation and engagement
    link.springer.com/article/10.1
    #AIEd #Education #EdTech

  11. Designing a mobile chatbot-based learning journaling system for intrinsic motivation and engagement
    link.springer.com/article/10.1
    #AIEd #Education #EdTech

  12. The LLM Fallacy: Misattribution in AI-Assisted Cognitive Workflows
    arxiv.org/abs/2604.14807
    "a cognitive attribution error in which individuals misinterpret LLM-assisted outputs as evidence of their own independent competence, producing a systematic divergence between perceived and actual capability"
    #AIEd #psy #hci #LLM

  13. The LLM Fallacy: Misattribution in AI-Assisted Cognitive Workflows
    arxiv.org/abs/2604.14807
    "a cognitive attribution error in which individuals misinterpret LLM-assisted outputs as evidence of their own independent competence, producing a systematic divergence between perceived and actual capability"
    #AIEd #psy #hci #LLM

  14. The LLM Fallacy: Misattribution in AI-Assisted Cognitive Workflows
    arxiv.org/abs/2604.14807
    "a cognitive attribution error in which individuals misinterpret LLM-assisted outputs as evidence of their own independent competence, producing a systematic divergence between perceived and actual capability"
    #AIEd #psy #hci #LLM

  15. The LLM Fallacy: Misattribution in AI-Assisted Cognitive Workflows
    arxiv.org/abs/2604.14807
    "a cognitive attribution error in which individuals misinterpret LLM-assisted outputs as evidence of their own independent competence, producing a systematic divergence between perceived and actual capability"
    #AIEd #psy #hci #LLM

  16. The LLM Fallacy: Misattribution in AI-Assisted Cognitive Workflows
    arxiv.org/abs/2604.14807
    "a cognitive attribution error in which individuals misinterpret LLM-assisted outputs as evidence of their own independent competence, producing a systematic divergence between perceived and actual capability"
    #AIEd #psy #hci #LLM

  17. SafeTutors: Benchmarking Pedagogical Safety in AI Tutoring Systems arxiv.org/abs/2603.17373 "risk is answer over-disclosure, misconception reinforcement, and the abdication of scaffolding" "multi-turn dialogue worsens behavior, with pedagogical failures rising from 17.7% to 77.8%." #AIEd #EdTech

    SafeTutors: Benchmarking Pedag...

  18. SafeTutors: Benchmarking Pedagogical Safety in AI Tutoring Systems
    arxiv.org/abs/2603.17373
    "the primary risk is not toxic content but the quiet erosion of learning through answer over-disclosure, misconception reinforcement, and the abdication of scaffolding"
    "We uncover that all models show broad harm; scale doesn't reliably help; and multi-turn dialogue worsens behavior, with pedagogical failures rising from 17.7% to 77.8%."
    #AIEd #EdTech

  19. SafeTutors: Benchmarking Pedagogical Safety in AI Tutoring Systems
    arxiv.org/abs/2603.17373
    "the primary risk is not toxic content but the quiet erosion of learning through answer over-disclosure, misconception reinforcement, and the abdication of scaffolding"
    "We uncover that all models show broad harm; scale doesn't reliably help; and multi-turn dialogue worsens behavior, with pedagogical failures rising from 17.7% to 77.8%."
    #AIEd #EdTech

  20. SafeTutors: Benchmarking Pedagogical Safety in AI Tutoring Systems
    arxiv.org/abs/2603.17373
    "the primary risk is not toxic content but the quiet erosion of learning through answer over-disclosure, misconception reinforcement, and the abdication of scaffolding"
    "We uncover that all models show broad harm; scale doesn't reliably help; and multi-turn dialogue worsens behavior, with pedagogical failures rising from 17.7% to 77.8%."
    #AIEd #EdTech

  21. SafeTutors: Benchmarking Pedagogical Safety in AI Tutoring Systems
    arxiv.org/abs/2603.17373
    "the primary risk is not toxic content but the quiet erosion of learning through answer over-disclosure, misconception reinforcement, and the abdication of scaffolding"
    "We uncover that all models show broad harm; scale doesn't reliably help; and multi-turn dialogue worsens behavior, with pedagogical failures rising from 17.7% to 77.8%."
    #AIEd #EdTech

  22. SafeTutors: Benchmarking Pedagogical Safety in AI Tutoring Systems
    arxiv.org/abs/2603.17373
    "the primary risk is not toxic content but the quiet erosion of learning through answer over-disclosure, misconception reinforcement, and the abdication of scaffolding"
    "We uncover that all models show broad harm; scale doesn't reliably help; and multi-turn dialogue worsens behavior, with pedagogical failures rising from 17.7% to 77.8%."
    #AIEd #EdTech

  23. EduQwen: Application-Driven Pedagogical Knowledge Optimization of Open-Source LLMs via Reinforcement Learning and Supervised Fine-Tuning
    arxiv.org/abs/2604.06385
    A fine-tuned open #LLM beats even Gemini on a #pedagogy benchmark. Unfortunately it doesn't appear to be released yet.
    #AIEd

  24. EduQwen: Application-Driven Pedagogical Knowledge Optimization of Open-Source LLMs via Reinforcement Learning and Supervised Fine-Tuning
    arxiv.org/abs/2604.06385
    A fine-tuned open #LLM beats even Gemini on a #pedagogy benchmark. Unfortunately it doesn't appear to be released yet.
    #AIEd

  25. EduQwen: Application-Driven Pedagogical Knowledge Optimization of Open-Source LLMs via Reinforcement Learning and Supervised Fine-Tuning
    arxiv.org/abs/2604.06385
    A fine-tuned open #LLM beats even Gemini on a #pedagogy benchmark. Unfortunately it doesn't appear to be released yet.
    #AIEd

  26. EduQwen: Application-Driven Pedagogical Knowledge Optimization of Open-Source LLMs via Reinforcement Learning and Supervised Fine-Tuning
    arxiv.org/abs/2604.06385
    A fine-tuned open #LLM beats even Gemini on a #pedagogy benchmark. Unfortunately it doesn't appear to be released yet.
    #AIEd

  27. EduQwen: Application-Driven Pedagogical Knowledge Optimization of Open-Source LLMs via Reinforcement Learning and Supervised Fine-Tuning
    arxiv.org/abs/2604.06385
    A fine-tuned open #LLM beats even Gemini on a #pedagogy benchmark. Unfortunately it doesn't appear to be released yet.
    #AIEd

  28. ISD-Agent-Bench: A Comprehensive Benchmark for Evaluating LLM-based Instructional Design Agents arxiv.org/abs/2602.10620 Code & data: github.com/codingchild2... Also: Pedagogy-R1: Pedagogical Reasoning Model and Educational Benchmark dl.acm.org/doi/10.1145/... #AIEd #LearningDesign #EdTech

    ISD-Agent-Bench: A Comprehensi...

  29. ISD-Agent-Bench: A Comprehensive Benchmark for Evaluating LLM-based Instructional Design Agents
    arxiv.org/abs/2602.10620
    Code & data: github.com/codingchild2424/isd
    "benchmark comprising 25,795 scenarios that combines 51 contextual variables across 5 categories with 33 ISD sub-steps derived from the ADDIE model."

    w/same author: Pedagogy-R1: Pedagogical Large Reasoning Model and Well-balanced Educational Benchmark dl.acm.org/doi/10.1145/3746252
    #AIEd #LearningDesign #AIevaluation #EdTech

  30. ISD-Agent-Bench: A Comprehensive Benchmark for Evaluating LLM-based Instructional Design Agents
    arxiv.org/abs/2602.10620
    Code & data: github.com/codingchild2424/isd
    "benchmark comprising 25,795 scenarios that combines 51 contextual variables across 5 categories with 33 ISD sub-steps derived from the ADDIE model."

    w/same author: Pedagogy-R1: Pedagogical Large Reasoning Model and Well-balanced Educational Benchmark dl.acm.org/doi/10.1145/3746252
    #AIEd #LearningDesign #AIevaluation #EdTech

  31. ISD-Agent-Bench: A Comprehensive Benchmark for Evaluating LLM-based Instructional Design Agents
    arxiv.org/abs/2602.10620
    Code & data: github.com/codingchild2424/isd
    "benchmark comprising 25,795 scenarios that combines 51 contextual variables across 5 categories with 33 ISD sub-steps derived from the ADDIE model."

    w/same author: Pedagogy-R1: Pedagogical Large Reasoning Model and Well-balanced Educational Benchmark dl.acm.org/doi/10.1145/3746252
    #AIEd #LearningDesign #AIevaluation #EdTech

  32. ISD-Agent-Bench: A Comprehensive Benchmark for Evaluating LLM-based Instructional Design Agents
    arxiv.org/abs/2602.10620
    Code & data: github.com/codingchild2424/isd
    "benchmark comprising 25,795 scenarios that combines 51 contextual variables across 5 categories with 33 ISD sub-steps derived from the ADDIE model."

    w/same author: Pedagogy-R1: Pedagogical Large Reasoning Model and Well-balanced Educational Benchmark dl.acm.org/doi/10.1145/3746252
    #AIEd #LearningDesign #AIevaluation #EdTech

  33. ISD-Agent-Bench: A Comprehensive Benchmark for Evaluating LLM-based Instructional Design Agents
    arxiv.org/abs/2602.10620
    Code & data: github.com/codingchild2424/isd
    "benchmark comprising 25,795 scenarios that combines 51 contextual variables across 5 categories with 33 ISD sub-steps derived from the ADDIE model."

    w/same author: Pedagogy-R1: Pedagogical Large Reasoning Model and Well-balanced Educational Benchmark dl.acm.org/doi/10.1145/3746252
    #AIEd #LearningDesign #AIevaluation #EdTech

  34. Knowledge graphs are useful representations for knowledge bases, #pkm, #AImemory systems, #GraphRAG, intelligent tutoring systems, etc., and usually implemented in graph databases. LadybugDB, a fork of the discontinued Kuzu, is a lightweight embedded (like SQLite) graph database: github.com/LadybugDB/ladybug
    Sample applications in development: github.com/inventivepotter/dot & github.com/tejzpr/Smriti-MCP
    See also Grafeo: github.com/GrafeoDB/grafeo
    #AIEd #AIEngineering #KnowledgeGraph #GraphDB #graphdatabase

  35. Knowledge graphs are useful representations for knowledge bases, #pkm, #AImemory systems, #GraphRAG, intelligent tutoring systems, etc., and usually implemented in graph databases. LadybugDB, a fork of the discontinued Kuzu, is a lightweight embedded (like SQLite) graph database: github.com/LadybugDB/ladybug
    Sample applications in development: github.com/inventivepotter/dot & github.com/tejzpr/Smriti-MCP
    See also Grafeo: github.com/GrafeoDB/grafeo
    #AIEd #AIEngineering #KnowledgeGraph #GraphDB #graphdatabase

  36. Knowledge graphs are useful representations for knowledge bases, #pkm, #AImemory systems, #GraphRAG, intelligent tutoring systems, etc., and usually implemented in graph databases. LadybugDB, a fork of the discontinued Kuzu, is a lightweight embedded (like SQLite) graph database: github.com/LadybugDB/ladybug
    Sample applications in development: github.com/inventivepotter/dot & github.com/tejzpr/Smriti-MCP
    See also Grafeo: github.com/GrafeoDB/grafeo
    #AIEd #AIEngineering #KnowledgeGraph #GraphDB #graphdatabase

  37. Knowledge graphs are useful representations for knowledge bases, #pkm, #AImemory systems, #GraphRAG, intelligent tutoring systems, etc., and usually implemented in graph databases. LadybugDB, a fork of the discontinued Kuzu, is a lightweight embedded (like SQLite) graph database: github.com/LadybugDB/ladybug
    Sample applications in development: github.com/inventivepotter/dot & github.com/tejzpr/Smriti-MCP
    See also Grafeo: github.com/GrafeoDB/grafeo
    #AIEd #AIEngineering #KnowledgeGraph #GraphDB #graphdatabase

  38. Knowledge graphs are useful representations for knowledge bases, #pkm, #AImemory systems, #GraphRAG, intelligent tutoring systems, etc., and usually implemented in graph databases. LadybugDB, a fork of the discontinued Kuzu, is a lightweight embedded (like SQLite) graph database: github.com/LadybugDB/ladybug
    Sample applications in development: github.com/inventivepotter/dot & github.com/tejzpr/Smriti-MCP
    See also Grafeo: github.com/GrafeoDB/grafeo
    #AIEd #AIEngineering #KnowledgeGraph #GraphDB #graphdatabase