Podcasts about Wikidata

Free knowledge database project

  • 68PODCASTS
  • 122EPISODES
  • 50mAVG DURATION
  • 1MONTHLY NEW EPISODE
  • Nov 4, 2025LATEST
Wikidata

POPULARITY

20172018201920202021202220232024


Best podcasts about Wikidata

Latest podcast episodes about Wikidata

SEO Is Not That Hard
Entities Part 3 : The Knowledge Graph

SEO Is Not That Hard

Play Episode Listen Later Oct 10, 2025 10:59 Transcription Available


Send us a textMost brands still try to “tell” Google who they are. We show how Google actually decides: by stitching together a ledger of facts from your site, LinkedIn, Wikipedia, news articles, and structured data—then trusting only what aligns. This is the Knowledge Graph at work, and it's quietly steering whether you earn a knowledge panel, sitelinks, and richer visibility across search.We break down the four streams feeding the graph—public web pages, licensed datasets, human‑edited knowledge bases like Wikidata, and direct owner signals via schema.org—and explain how each contributes to a confidence score for your entity. If your about page says Jane Doe is CEO but LinkedIn shows John Smith, the score drops and your brand becomes ambiguous. If your website, LinkedIn, reputable press, and Wikidata all agree, trust rises and your facts become “truth” in search.From there, we get specific about what you can control. Use schema.org to describe your organisation, people, products, and identifiers in clear, machine‑readable terms. Link out with sameAs to authoritative profiles so Google can triangulate identity. Audit your knowledge panel as a live diagnostic: check logos, dates, roles, and categories, and chase down any mismatch to the original source. Treat digital PR and reputation management as part of technical SEO—because today they are.By the end, you'll have a practical checklist for entity hygiene that helps you earn and keep a clean knowledge panel, avoid costly confusion, and unlock higher‑trust features across the results page. If this helped clarify how entities power modern SEO, subscribe, share with a colleague, and leave a quick review with one takeaway you'll act on next.SEO Is Not That Hard is hosted by Edd Dawson and brought to you by KeywordsPeopleUse.com Help feed the algorithm and leave a review at ratethispodcast.com/seo You can get your free copy of my 101 Quick SEO Tips at: https://seotips.edddawson.com/101-quick-seo-tipsTo get a personal no-obligation demo of how KeywordsPeopleUse could help you boost your SEO and get a 7 day FREE trial of our Standard Plan book a demo with me nowSee Edd's personal site at edddawson.comAsk me a question and get on the show Click here to record a questionFind Edd on Linkedin, Bluesky & TwitterFind KeywordsPeopleUse on Twitter @kwds_ppl_use"Werq" Kevin MacLeod (incompetech.com)Licensed under Creative Commons: By Attribution 4.0 Licensehttp://creativecommons.org/licenses/by/4.0/

SEO Is Not That Hard
Entities Part 2 : How Machines Learn To Read

SEO Is Not That Hard

Play Episode Listen Later Oct 8, 2025 11:59 Transcription Available


Send us a textKeywords don't tell the whole story—entities do. We take you inside the three-step process machines use to read your content like a detective at a crime scene: highlighting potential entities, using context to resolve ambiguity, and linking each mention to a unique identifier in a global knowledge base. By the end, you'll see why “Jordan” only makes sense when surrounded by the right clues—and how to present those clues so search engines and AIs make the right call every time.We start with named entity recognition, the digital highlighter that picks out people, organisations, products, places, and dates across unstructured text. Then we move to entity disambiguation, where context—co-occurring teams, locations, or concepts—guides the system to the correct meaning. Finally, we close with entity linking, the moment a string becomes a node with a library card in Wikipedia or Wikidata. That linkage is the bridge into Google's Knowledge Graph, powering features like knowledge panels and richer, more confident results.Along the way, we dig into why Wikipedia and Wikidata matter far beyond vanity. Accurate, well-sourced entries create a feedback loop that improves how machines understand your brand, your founders, and your products. If you don't meet notability yet, don't force it; build authority elsewhere with consistent profiles, structured data, and content that names and connects related entities. We also share a simple action: search for your brand, founder, and main product on Wikipedia and Wikidata and assess accuracy. Want more like this? Follow the show, share it with a colleague, and leave a review so we can help more teams make sense of entity-first SEO.SEO Is Not That Hard is hosted by Edd Dawson and brought to you by KeywordsPeopleUse.com Help feed the algorithm and leave a review at ratethispodcast.com/seo You can get your free copy of my 101 Quick SEO Tips at: https://seotips.edddawson.com/101-quick-seo-tipsTo get a personal no-obligation demo of how KeywordsPeopleUse could help you boost your SEO and get a 7 day FREE trial of our Standard Plan book a demo with me nowSee Edd's personal site at edddawson.comAsk me a question and get on the show Click here to record a questionFind Edd on Linkedin, Bluesky & TwitterFind KeywordsPeopleUse on Twitter @kwds_ppl_use"Werq" Kevin MacLeod (incompetech.com)Licensed under Creative Commons: By Attribution 4.0 Licensehttp://creativecommons.org/licenses/by/4.0/

The Tech Blog Writer Podcast
3391: What Wikidata Reveals About the Good Side of the Internet

The Tech Blog Writer Podcast

Play Episode Listen Later Aug 19, 2025 19:09


When most people think of Wikipedia, they picture an endless scroll of human-readable pages. But there's another side to this ecosystem, one designed not just for people but also for machines. It's called Wikidata, and if you haven't heard of it, that's exactly why this conversation matters. In this episode of Tech Talks Daily, I'm joined by Lydia Pintscher, Wikidata Portfolio Manager at Wikimedia Deutschland, for a deep look into how structured, open data is quietly powering civic tech, cultural preservation, and knowledge equity across the globe. Wikidata is the backbone that helps turn static knowledge into something living, adaptable, and scalable. With over 117 million items, 1.65 billion semantic statements, and more than 2.34 billion edits, it's become one of the largest collaborative datasets in the world. But it's not just the size that makes it impressive. It's what people are doing with it. Lydia shares how volunteers and developers are building tools for everything from investigative journalism to public libraries, all without needing deep pockets or proprietary infrastructure. This isn't big tech. It's a global, grassroots movement making open data work for the public good. We explore how tools like Toolforge and the Wikidata Query Service lower the barrier to entry, allowing civil society groups to build sophisticated applications that would otherwise be out of reach. Whether it's helping connect citizens to government services or preserving disappearing languages, the use cases are multiplying fast. Lydia also reflects on how Wikidata fosters a sense of purpose for contributors, offering a rare example of what many call the good internet, where collaboration outweighs competition and building something meaningful beats chasing virality. If you're curious about where open knowledge is headed, how structured data can be a force for social impact, or why Wikidata might be the most important project you've never fully explored, this episode offers a window into a future where machines help humans build something better, together.

Open||Source||Data
Building Open-Source LLMs with Philosophy | Anastasia Stasenko

Open||Source||Data

Play Episode Listen Later Jun 3, 2025 57:45


Join Charna Parkey as she welcomes Anastasia Stasenko, CEO and co-founder of pleias, through her unique journey from philosophy to building open-source, energy-efficient LLMs. Discover how pleias is revolutionizing the AI landscape by training models exclusively on open data and establishing a precedent for ethical and socially acceptable AI. Learn about the challenges and opportunities in creating multilingual models and contributing back to the open-source community. QUOTES[00:00:00] Introducing Anastasia and pleias[00:02:00] From Philosophy to AI[00:06:00] The Problem of Generic Models[00:10:00] Open Weights vs. Open Source vs. Open Science[00:14:00] Why Open Data Matters[00:18:00] High-Quality, Specialized Models[00:22:00] Multilingual Challenges[00:26:00] Global Inclusion Requires Small Models[00:30:00] Using and Contributing to Wikidata[00:38:00] The Future: Specialized Models[00:48:00] Advice for Newcomers[00:54:00] Cultural Sensitivity and Data Representation[00:50:00] Leo's Takeaways[00:52:00] Charna on Ethical, Verifiable AI[00:54:00] Representation vs. Exclusion[00:56:00] Letting People Be More Human[00:57:30] Applied, Transformative AIQUOTESCharna:"If you didn't make it represented in the data, then we're leaving another culture behind... So which one are you wanting to do, misrepresent them or just completely leave them behind from this technical revolution?"Anastasia:"The real issue now is that the lack of diversity in the current AI labs leads to the situation where all LLMs look alike."Anastasia:"Being able to design, to find, and also to create the appropriate data mix for large language models is something that we shouldn't really forget about when we talk about the success of what large language models are."

Sustain
Episode 260: Robert Douglass and contributing as a corporation to OSS

Sustain

Play Episode Listen Later Dec 13, 2024 32:55


Guest Robert Douglass Panelist Richard Littauer | Abby Cabunoc Mayes Show Notes In this episode of Sustain, hosts Richard Littauer and Abby Cabunoc Mayes speak with Robert Douglass, Entrepreneur in Residence at Open Strategy Partners, to delve into sustaining open source projects. They explore Robert's extensive history with Drupal, the role of Open Strategy Partners, and the innovative Drupal Certified Partner Program designed to address the maker-taker dilemma in open source. The episode also covers the recently launched RFP templates aimed at promoting open source software and certified partners. Robert shares insights on gamification, the economic aspects of contributing to Drupal, and future initiatives to ensure the continued sustainability of open source projects. Hit download now to hear more! [00:01:49] Robert shares his background in the Drupal ecosystem and his involvement with Open Strategy Partners, which provides strategic content marketing for B2B tech companies focusing on open source. [00:02:43] Robert explains Open Strategy Partners' focus on supporting open source projects and mentions clients like DDEV and TYPO3. [00:04:06] Richard and Robert discuss what it means to be an entrepreneur in residence, with Robert explaining his role in developing new products for Open Strategy Partners and the books he has written. [00:05:52] Robert reflects on the early days of Drupal and the challenges in making open source sustainable. He notes how the community was initially driven by passion, with few paid opportunities. [00:08:05] Robert introduces the Drupal Certified Partner Program, a system for supporting Drupal sustainability by encouraging companies to contribute both time and money. [00:10:03] The conversation covers how Drupal's contribution system gamifies the support companies provide to the ecosystem. Companies can earn contribution credits, which are visible on Drupal.org and benefit their reputation. [00:15:41] Abby asks about the potential downsides of gamification, especially regarding diversity. Robert explains how placing the system at the company level may mitigate some negative impacts. [00:18:17] Richard inquires about the financial structure of the Drupal Certified Partner Program. Robert clarifies that the funds collected support the Drupal Association's core mission, including maintaining Drupal.org and organizing events. [00:21:33] Robert discusses the development of RFP (Request for Proposal) templates to encourage companies to consider certified open source providers, explaining how this initiative promotes sustainability in the ecosystem. [00:25:56] Robert describes how the RFP templates allow purpose-driven organizations to incorporate open source values in their procurement process, aligning with their missions. [00:27:00] Robert invites listeners to explore and utilize the RFP templates, which are available under a Creative Commons Zero license, encouraging others to adapt and improve them. [00:29:47] Find out where you can follow Robert and his work online. Quotes [00:08:57] “Open Source is like a free puppy.” Spotlight [00:30:30] Abby's spotlight is Common Sort thrift shop in Toronto. [00:30:52] Richard's spotlight is Wikidata. [00:31:21] Robert's spotlight is Chad Whitacre and Sentry. Links SustainOSS (https://sustainoss.org/) podcast@sustainoss.org (mailto:podcast@sustainoss.org) richard@sustainoss.org (mailto:richard@sustainoss.org) SustainOSS Discourse (https://discourse.sustainoss.org/) SustainOSS Mastodon (https://mastodon.social/tags/sustainoss) Open Collective-SustainOSS (Contribute) (https://opencollective.com/sustainoss) Richard Littauer Socials (https://www.burntfen.com/2023-05-30/socials) Abby Cabunoc Mayes X (https://x.com/abbycabs?lang=en) Robert Douglass LinkedIn (https://www.linkedin.com/in/roberttdouglass/) Open Strategy Partners (https://openstrategypartners.com/) Open Strategy Partners Blog (https://openstrategypartners.com/blog/) Building Online Communities with Drupal, phpBB, and WordPress by Robert Douglass, Mike Little, Jared W. Smith (https://www.drupal.org/node/1850002) Drupal Certified Partner Program (https://www.drupal.org/association/become-a-drupal-certified-partner) Drupal (https://www.drupal.org/) How to Write an RFP for Open Source Solutions: Featuring Drupal Certified Partners (https://www.drupal.org/association/blog/how-to-write-an-rfp-for-open-source-solutions-featuring-drupal-certified-partners) OSP: Supporting Drupal Certified Partners (https://openstrategypartners.com/blog/osp-supporting-drupal-certified-partners/) Sustain Podcast-Episode 148: Ali Nehzat of thanks.dev and OSS Funding (https://podcast.sustainoss.org/148) Common Sort (https://commonsort.com/) Wikidata (https://www.wikidata.org/wiki/Wikidata:Main_Page) Chad Whitacre LinkedIn (https://www.linkedin.com/in/chadwhitacre/) Sentry (https://sentry.io/welcome/) Credits Produced by Richard Littauer (https://www.burntfen.com/) Edited by Paul M. Bahr at Peachtree Sound (https://www.peachtreesound.com/) Show notes by DeAnn Bahr Peachtree Sound (https://www.peachtreesound.com/) Special Guest: Robert Douglass.

#arthistoCast – der Podcast zur Digitalen Kunstgeschichte
Folge 16: Wissensgerechtigkeit auf Wikipedia: Kunsthistoriker*innen gestalten mit

#arthistoCast – der Podcast zur Digitalen Kunstgeschichte

Play Episode Listen Later Nov 6, 2024 48:22


In dieser Folge spricht Jacqueline Klusik-Eckert mit den Kunsthistorikerinnen Anna Gnyp und Maria Merseburger über das Verhältnis von Wikimedia und der Kunstgeschichte. Gemeinsam diskutieren sie, wie Wikipedia und Wikidata inzwischen zu wertvollen Ressourcen für kunsthistorische Forschung geworden sind und warum die aktive Mitgestaltung dieser Plattformen durch Fachwissenschaftler*innen unter dem Aspekt der Wissensgerechtigkeit wichtig ist. Dabei wird auch die Arbeit der AG Kuwiki vorgestellt, die mit mehreren Projekten die Sichtbarkeit kunsthistorischen Wissens auf Wikipedia fördert: Das „Living Handbook“ bietet eine Einführung in die Wikipedia-Arbeit für Kunsthistoriker. „Wikipedia in der Lehre“ zielt darauf ab, Studierende frühzeitig für die Plattform zu sensibilisieren und aktiv einzubinden. Und „Kuwiki Loves Monuments, too“ fördert die Dokumentation und Verbreitung von Bildern zu Denkmälern und Kulturgütern. Ein wichtiges Anliegen ist dabei die Wissensgerechtigkeit, um mehr Diversität auf Wikipedia und Wikimedia Commons zu erreichen.Das Gespräch beleuchtet auch die wachsende Bedeutung von Wikidata als datenbankgestützte Ressource, die zunehmend in digitalen kunsthistorischen Projekten genutzt wird. Anna und Maria zeigen auf, wie Museen, Archive und Bibliotheken von Wikidata und Wikimedia Commons profitieren können, um ihre Bestände öffentlich zugänglich zu machen und neue Vernetzungen zu schaffen. Abschließend plädieren sie für stärkere Kooperationen und „Best Practice“-Beispiele, die die Arbeit mit Wikimedia-Projekten in der Kunstwissenschaft festigen und bereichern können.Anna Gnyp, ist seit knapp zwei Jahren Mitglied der Arbeitsgemeinschaft. Aktuell ist sie Wissenschaftlerin im Datenkompetenzzentrum Sammlungen, Objekte, Datenkompetenz an der Humboldt-Universität Berlin. Das ist ein Verbundprojekt zum Aufbau eines Datenkompetenzzentrums für wissenschaftliche Universitätssammlungen.Dr. Maria Merseburger, ist seit Beginn im der AG KUwiki, hier unter dem Namen Karatecoop und aktuell Wissenschaftlerin am Museum für Kommunikation in Berlin.Begleitmaterial zu den Folgen findest du auf der Homepage unter https://www.arthistoricum.net/themen/podcasts/arthistocast.Alle Folgen des Podcasts werden bei heidICON mit Metadaten und persistentem Identifier gespeichert. Die Folgen haben die Creative-Commons-Lizenz CC BY 4.0 und können heruntergeladen werden. Du findest sie unter https://doi.org/10.11588/heidicon/1738702.Bei Fragen, Anregungen, Kritik und gerne auch Lob kannst du uns gerne per Mail kontaktieren unter podcast@digitale-kunstgeschichte.de.

Somewhere on Earth: The Global Tech Podcast
Could making Wikidata 'human' readable lead to better AI?

Somewhere on Earth: The Global Tech Podcast

Play Episode Listen Later Oct 10, 2024 31:41


Could making Wikidata 'human' readable lead to better AI? A new project is underway to allow Large Language Models (LLMs) to read Wikidata. The data is currently structured in a way that's machine readable, but LLMs read data more like humans than machines, meaning this vast amount of human curated, high quality data isn't accessible to this type of AI. By allowing access to Wikidata, LLMs could become more reliable. Ania spoke to Lydia Pintscher, the Portfolio Lead Product Manager at Wikidata Deutschland, to learn more about these developments. Most news websites block AI Chatbots Two thirds of high quality news websites block AI chatbots from accessing their information, according to a report by the misinformation monitoring organisation NewsGuard. This means that some of the world's most popular AI chatbots could be collecting data on misinformation from low quality news sources and even conspiracy and hoax sites. The Enterprise Editor at NewsGuard is Jack Brewster and he is on the show to explain their findings. The programme is presented by Gareth Mitchell and the studio expert is Ania Lichtarowicz. More on this week's stories: Wikidata and Artificial Intelligence: Simplified Access to Open Data for Open-Source Projects AI Chatbots Are Blocked by 67% of Top News Sites, Relying Instead on Low-Quality Sources Support the show Editor: Ania Lichtarowicz Production Manager: Liz Tuohy Recording and audio editing : Lansons | Team Farner For new episodes, subscribe wherever you get your podcasts or via this link: https://www.buzzsprout.com/2265960/supporters/new Follow us on all the socials: Join our Facebook group Instagram Twitter/X If you like Somewhere on Earth, please rate and review it on Apple Podcasts Contact us by email: hello@somewhereonearth.co Send us a voice note: via WhatsApp: +44 7486 329 484 Find a Story + Make it News = Change the World Learn more about your ad choices. Visit megaphone.fm/adchoices

Somewhere on Earth: The Global Tech Podcast
Could making Wikidata 'human' readable lead to better AI?

Somewhere on Earth: The Global Tech Podcast

Play Episode Listen Later Oct 8, 2024 27:26 Transcription Available


Could making Wikidata 'human' readable lead to better AI? A new project is underway to allow Large Language Models (LLMs) to read Wikidata. The data is currently structured in a way that's machine readable, but LLMs read data more like humans than machines, meaning this vast amount of human curated, high quality data isn't accessible to this type of AI. By allowing access to Wikidata, LLMs could become more reliable. Ania spoke to Lydia Pintscher, the Portfolio Lead Product Manager at Wikidata Deutschland, to learn more about these developments. Most news websites block AI ChatbotsTwo thirds of high quality news websites block AI chatbots from accessing their information, according to a report by the misinformation monitoring organisation NewsGuard. This means that some of the world's most popular AI chatbots could be collecting data on misinformation from low quality news sources and even conspiracy and hoax sites. The Enterprise Editor at NewsGuard is Jack Brewster and he is on the show to explain their findings.The programme is presented by Gareth Mitchell and the studio expert is Ania Lichtarowicz.More on this week's stories: Wikidata and Artificial Intelligence: Simplified Access to Open Data for Open-Source ProjectsAI Chatbots Are Blocked by 67% of Top News Sites, Relying Instead on Low-Quality SourcesSupport the showEditor: Ania LichtarowiczProduction Manager: Liz Tuohy Recording and audio editing : Lansons | Team Farner For new episodes, subscribe wherever you get your podcasts or via this link:https://www.buzzsprout.com/2265960/supporters/newFollow us on all the socials: Join our Facebook group Instagram Twitter/X If you like Somewhere on Earth, please rate and review it on Apple PodcastsContact us by email: hello@somewhereonearth.coSend us a voice note: via WhatsApp: +44 7486 329 484Find a Story + Make it News = Change the World

Somewhere on Earth: The Global Tech Podcast

Subscriber-only episodeCould making Wikidata human readable lead to better AI? A new project is underway to allow Large Language Models (LLMs) to read Wikidata. The data is currently structured in a way that's machine readable, but LLMs read data more like humans than machines, meaning this vast amount of human curated, high quality data isn't accessible to this type of AI. By allowing access to Wikidata, LLMs could become more reliable. Ania spoke to Lydia Pintscher, the Portfolio Lead Product Manager at Wikidata Deutschland, to learn more about these developments. Most news websites block AI ChatbotsTwo thirds of high quality news websites block AI chatbots from accessing their information, according to a report by the misinformation monitoring organisation NewsGuard. This means that some of the world's most popular AI chatbots could be collecting data on misinformation from low quality news sources and even conspiracy and hoax sites. The Enterprise Editor at NewsGuard is Jack Brewster and he is on the show to explain their findings.The programme is presented by Gareth Mitchell and the studio expert is Ania Lichtarowicz.More on this week's stories: Wikidata and Artificial Intelligence: Simplified Access to Open Data for Open-Source ProjectsAI Chatbots Are Blocked by 67% of Top News Sites, Relying Instead on Low-Quality SourcesEditor: Ania LichtarowiczProduction Manager: Liz Tuohy Recording and audio editing : Lansons | Team Farner For new episodes, subscribe wherever you get your podcasts or via this link:https://www.buzzsprout.com/2265960/supporters/newFollow us on all the socials: Join our Facebook group Instagram Twitter/X If you like Somewhere on Earth, please rate and review it on Apple PodcastsContact us by email: hello@somewhereonearth.coSend us a voice note: via WhatsApp: +44 7486 329 484Find a Story + Make it News = Change the World

Information Morning from CBC Radio Nova Scotia (Highlights)
Wikipedia edit-a-thon to improve Mi'kmaw content online

Information Morning from CBC Radio Nova Scotia (Highlights)

Play Episode Listen Later Aug 2, 2024 7:26


During National Indigenous History Month, Dalhousie University Libraries hosted an edit-a-thon to improve Wikipedia and Wikidata content related to Mi'kmaw people and Mi'kma'ki. A total of 19 Wikipedia articles were edited, 50 references added, and more than 3,300 words were contributed. One of the organizers fills us in. 

Pojačalo
EP 258: Goran S. Milovanović, DataKolektiv & Smartocto - Pojačalo podcast

Pojačalo

Play Episode Listen Later Mar 17, 2024 156:00


"Uvek sam radio samo ono što sam voleo i što me ložilo." Gost Ivana Minića u 258. epizodi Pojačala je dr Goran S. Milovanović, doktor psiholoških nauka koji nije pošao putem koji obično asociramo sa ovim poljem - psiholog, psihijatar ili psihoterapeut, već se specijalizovao za kognitivne i podatkovne nauke i machine learning. Polja koja zahtevaju duboko razumevanje ljudskog uma i razmišljanja, ali takođe i visok nivo tehnološkog znanja i stručnosti. Karijerni put dr Milovanovića mogao bi sam po sebi ispuniti tri epizode ove dužine, ali ova epizoda posvećena je njegovoj mladosti, priči o tome šta ga je stavilo na taj put, radu u Wikidata, najvećoj bazi znanja na svetu kao i kompaniji Smartocto koja se bavi prediktivnom analitikom u digitalnim medijima. Teme u epizodi: - Uvod - Početak razgovora - Kad porastem biću... - Dobri stari dani - Fakultetski dani - Fundamentalna nauka - Big data analytics - Kognitivne pristrasnosti - Wikidata I Smartocto - Problem digitalnih medija - Data kolektiv Realizacija Pojačalo podkasta ne bi bila moguća bez naših izuzetnih partnera: - Kompanija Epson koja je vodeći svetski proizvođač projektora i štampača za sve namene: https://www.epson.rs/sr_RS - Kompanija Orion telekom provajtera najbrže internet infrastrukture u Srbiji sa preko 30 godina iskustva: https://oriontelekom.rs Podržite nas na BuyMeACoffee: https://bit.ly/3uSBmoa Pročitajte transkript ove epizode: https://bit.ly/3PrfigS Posetite naš sajt i prijavite se na našu mailing listu: http://bit.ly/2LUKSBG Prijavite se na naš YouTube kanal: http://bit.ly/2Rgnu7o Pratite Pojačalo na društvenim mrežama: Facebook: http://bit.ly/2FfwqCR Twitter: http://bit.ly/2CVZoGr Instagram: http://bit.ly/2RzGHjN

Glitterbrains
010 - Monogames Lesen

Glitterbrains

Play Episode Listen Later Jan 20, 2024 119:31


In dieser ersten Folge des Jahres 2024 haben Letty und hukl wieder einen äußerst bunten Themenblumenstrauß für Euch mitgebracht. Schuhe, Game Engines, Trinkwasser und Osmose, Wikidata und ChatGPT und noch Vieles mehr findet den Weg in Eure Ohren und Synapsen. Bitte knacken Sie mit die Synapsen! Die Mixtape Playlists findet Ihr in den Shownotes (nicht 100% weil nicht Alles streambar).

Computer und Kommunikation Beiträge - Deutschlandfunk
Datenspende: Wikidata soll die Qualität von Large Language Modellen verbessern

Computer und Kommunikation Beiträge - Deutschlandfunk

Play Episode Listen Later Oct 28, 2023 6:56


Kloiber, Manfredwww.deutschlandfunk.de, Forschung aktuellDirekter Link zur Audiodatei

Pizzel Podcast
Pizzel Ep. 99 - Q177

Pizzel Podcast

Play Episode Listen Later Oct 12, 2023 67:02


Pedro se muda y nos cuenta de primera mano sus experiencias con fletes japoneses y después nos pone al día sobre las grandes ambiciones de Wikipedia y el proyecto Wikidata. Por último Javier le escribe una carta de amor a uno de sus músicos favoritos, Louis Cole, en forma de sección de podcast.

Escola Mobile. Biznes masz w kieszeni
Wikipedia: edukacja i technologia - Natalia Ćwik (CEO Wikimedia Polska) EM #156

Escola Mobile. Biznes masz w kieszeni

Play Episode Listen Later Sep 20, 2023 52:37


Pamiętasz jak pierwszy raz skorzystałeś z Wikipedii? A może korzystanie z Wiki jest już tak naturalne, że nawet nie zastanawiasz się nad tym, skąd biorą się hasła w tym projekcie? W tym odcinku podcastu gościmy CEO Wikimedia Polska. Natalia wespół z ogromem zaangażowanych ludzi buduje największy projekt społeczny w historii ludzkości. Stowarzyszenie Wikimedia Polska działa na rzecz powszechnego dostępu do wiedzy. Wspiera i promuje Wikipedię i jej projekty siostrzane (projekty Wikimedia). Jest niezależnym partnerem Wikimedia Foundation. Projekt z Wiki w nazwie to coś więcej niż pierwsze hasło w wyszukiwarce. Porozmawiamy o wolontariuszach, czyli wikipedystach, oraz szerokiej współpracy na rzecz edukacji. Uniwersytety, szkoły, firmy, startupy, galerie i biblioteki. Oraz każdy, kto chciałby współtworzyć Wikipedię i upowszechniać edukację. Wikipedia wyznacza trendy, buduje zasób edukacyjny dzięki zapleczu technologicznemu i energii ludzi. Dzięki temu Wikipedia to trzeci największy dostawca informacji w Unii Europejskiej, bez reklam, bez potrzeby komercjalizowania, utrzymujący się wyłącznie z darowizn, od ludzi, którzy chcą wspierać dostęp do wolnej wiedzy, nieustannie się rozrastający, I to w gigantycznym tempie, istniejący w ponad 300 wersjach językowych I coraz bardziej wewnętrznie ze sobą połączony, skoordynowany. Jak wygląda Wikipedia od kuchni? Czy AI jest już w Wikipedii oraz jak naprawiać wandalizmy w Wikipedii? Przesłuchaj podcast i sprawdź, o czym mówimy w tym odcinku. Logo dźwiękowe Wikipedii https://commons.m.wikimedia.org/wiki/File:Wikimedia_Sonic_Logo_-_4-seconds.wav Patronite https://patronite.pl/Wikipedia.   Projekt Wikiszkoła https://wikiszkola.pl   Wikimedia Commons   https://commons.m.wikimedia.org/wiki/Strona_główna    Wikimedia Foundation to amerykańska organizacja non-profit założona w celu rozwoju Wikipedii i jej projektów siostrzanych (takich jak Wikimedia Commons, Wikidata, Wikisource, Wiktionary itd). Utrzymuje m.in. serwery, na których znajdują się wszystkie światowe wersje Wikipedii.  Wikipedia jest największym i najbardziej rozpoznawalnym projektem prowadzonym przez WMF.  WMF wspiera również rozwój innych językowych Wikipedii i projektów siostrzanych na całym świecie.  Wikimedia Polska jest niezależnym partnerem Wikimedia Foundation, mającym wyłączne prawo na terenie Polski do używania marki Wikipedia.  Wikimedia Polska wspiera rozwój polskiej Wikipedii i jej projektów siostrzanych (np. Wikisłownik, Wikiźrodła, Wikicytaty, Wikidane, Wikimedia Commons i inne).    1 - Intro (00:01:10) 2 - O gościni (00:01:31) 3 - Wikipedia i Wikimedia (00:03:39) 4 - Jak pracują wikipedyści (00:05:56) 5 - Czym jest dziś Wikipedia (00:12:38) 6 - Praca w NBP (00:20:49) 7 - Wikipedia wyznacza trendy (00:23:51) 8 - Dobór treści w Wikipedii (00:25:44) 9 - Wikipedia na mobile (00:28:58) 10 - Dostępność (00:32:21) 11 - Współpraca z dostawcami technologii (00:34:40) 12 - AI i Wikipedia (00:37:49) 13 - Wandalizmy na Wikipedii (00:41:49) 14 - Przyszłość Wikipedii (00:46:53) 15 - Outro (00:50:22)   Muzyka: Kevin MacLeod Werq Kevin MacLeod (incompetech.com) Licensed under Creative Commons: By Attribution 4.0 License/mix by Jedrzej Paulus https://creativecommons.org/licenses/by/4.0/ Oceń nasz podcast na Apple Podcasts: https://bit.ly/EscolaMobileIT

#arthistoCast – der Podcast zur Digitalen Kunstgeschichte
Folge 6: Normdaten in der Kunstgeschichte

#arthistoCast – der Podcast zur Digitalen Kunstgeschichte

Play Episode Listen Later Sep 6, 2023 56:55


In dieser Folge spricht Jacqueline Klusik-Eckert mit Angela Kailus M.A. und Julia Rössel M.A. über die Rolle von Normdaten in der kunsthistorischen Forschung und Praxis. Der Ursprung von Normdaten hängt mit einem bibliothekarischen Systematisierungsbestrebungen in den 1970er Jahren zusammen. Wie hat sich der Umgang und die Konzepte von Normdaten im Zuge der Digitalisierung verändert? Mit einem Blick hinter die Kulissen der Gemeinsamen Normdatei (kurz GND) werden die Zusammenhänge von Identifikationsnummer und den dahinterliegenden Informationen erklärt. Welchen Mehrwert für die eigenen Daten erzielt man durch die Verwendung von Normdaten? Für welche Begriffe bzw. Entitäten gibt es Normdaten? Wo findet man sie? Woher kommt dieses Wissen und wie muss man mit dem Normdatensatz umgehen? Wir sprechen auch über den Unterschied eines institutionell gepflegten und autorisierten Normdatensatzes (GND über die Deutsche Nationalbibliothek) im Vergleich zu crowd-based Normdatensätzen (Wikidata).Darüber hinaus stellt sich die Frage, inwieweit die Verwendung von Normdaten bereits Einzug in das Fach Kunstgeschichte gehalten hat.Wir beleuchten die Herausforderung für sammelnde Institutionen bei der Erfassung von Objekten und der Anreicherung der Sammlungsdaten mit Normdaten. Welche Standards helfen bei der Erfassung und wofür soll man den Aufwand mit Normdaten überhaupt betreiben?Dabei nehmen wir unterschiedliche Szenarien im Datenlebenszyklus unter die Lupe.Wo begegnen wir als Forscher*innen diesen Daten, wie können wir sie nachnutzen und welche Verantwortung haben wir selbst als Produzent*innen von Forschungsdaten, wenn es um die Anreicherung der eigenen Daten mit Normdaten geht?Angela Kailus M.A. ist stellvertretende Leiterin des Deutschen Dokumentationszentrums für Kunstgeschichte – Bildarchiv Foto Marburg, Philipps-Universität Marburg und Ansprechperson bei NFDI4Culture im Arbeitsbereich Standardisierung und Datenqualität.Julia Rössel M.A. ist Kunsthistorikerin und Mitarbeiterin an der Deutschen Digitalen Bibliothek, Fachstelle Denkmalpflege, DDK, Marburg . Neben ihrer Promotion über „Wechsel der Mediensysteme – Graphische Sammlung und ihre digitale Übersetzung“ hat sie sich in den Bereichen Digitalisierung und Museum, Datenqualität und Standards spezialisiert.Begleitmaterial zu den Folgen findest du auf der Homepage unter https://www.arthistoricum.net/themen/podcasts/arthistocastAlle Folgen des Podcasts werden bei heidICON mit Metadaten und persistentem Identifier gespeichert. Die Folgen haben die Creative-Commons-Lizenz CC BY 4.0 und können heruntergeladen werden. Du findest sie unterhttps://doi.org/10.11588/heidicon/1738702 Bei Fragen, Anregungen, Kritik und gerne auch Lob kannst du gerne per Mail an uns schicken unterpodcast@digitale-kunstgeschichte.de

Wikipediapodden
Avsnitt 230 – nu ringer vectorklockan

Wikipediapodden

Play Episode Listen Later Sep 6, 2023 21:26


Vi klurar på hur andra källor återanvänder data från Wikipedia och Wikidata, hur referenslistor kan läggas till en masse, om den stora Vektoromställningen som nyss annonserades och ljud i rad.

On the Media
The Lasting Impact of the Library of Alexandria

On the Media

Play Episode Listen Later Aug 16, 2023 16:24


In the first half of the last school year, PEN America has recorded almost 900 different books pulled from library shelves across the country. As long as libraries have existed, people have tried to police what goes in them. The burning of the Library of Alexandria is a metaphor that gets invoked any time we lose access to a treasure trove of books. But for centuries it has also inspired scientists and inventors, philosophers and programmers to dream about creating an ideal library, one that provides access to all the knowledge in the world. OTM producer Molly Schwartz goes to a birthday party for Wikidata at the Brooklyn Public Library, where she talks to Wikimedia New York City president Richard Knipel, Wikimedia software engineer James Forrester, and long-time Wikipedia editor Jim Henderson about how the free online encyclopedia has made strides toward providing knowledge to the sum of human knowledge. She also speaks with library historian Alex Wright, author of the book Glut: Mastering Information Through the Ages, and software engineering consultant Gyula Lakatos, creator of the Library of Alexandria application suite, about the history of universal library projects and what keeps the dream alive. 

On the Media
The Lasting Impact of the Library of Alexandria

On the Media

Play Episode Listen Later Aug 16, 2023 16:21


In the first half of the last school year, PEN America has recorded almost 900 different books pulled from library shelves across the country. As long as libraries have existed, people have tried to police what goes in them. The burning of the Library of Alexandria is a metaphor that gets invoked any time we lose access to a treasure trove of books. But for centuries it has also inspired scientists and inventors, philosophers and programmers to dream about creating an ideal library, one that provides access to all the knowledge in the world. OTM producer Molly Schwartz goes to a birthday party for Wikidata at the Brooklyn Public Library, where she talks to Wikimedia New York City president Richard Knipel, Wikimedia software engineer James Forrester, and long-time Wikipedia editor Jim Henderson about how the free online encyclopedia has made strides toward providing knowledge to the sum of human knowledge. She also speaks with library historian Alex Wright, author of the book Glut: Mastering Information Through the Ages, and software engineering consultant Gyula Lakatos, creator of the Library of Alexandria application suite, about the history of universal library projects and what keeps the dream alive. 

The Nonlinear Library
AF - An LLM-based “exemplary actor” by Roman Leventov

The Nonlinear Library

Play Episode Listen Later May 29, 2023 24:24


Welcome to The Nonlinear Library, where we use Text-to-Speech software to convert the best writing from the Rationalist and EA communities into audio. This is: An LLM-based “exemplary actor”, published by Roman Leventov on May 29, 2023 on The AI Alignment Forum. Into and summary This post is the second section of "Aligning an H-JEPA agent via training on the outputs of an LLM-based "exemplary actor", posted separately because I think it could warrant a separate discussion, largely independent of the discussion of H-JEPA agent with GFlowNet actors. Here's the summary of this post, copied from the "Overview" section of the main article: In section 2, I describe the “exemplary actor”, an LMCA (language model cognitive architecture) that takes a simple, “brute force” approach to alignment: a powerful LLM (think GPT-5/6 level, with a vast, or quasi-unlimited context) is given a list of “approved” textbooks on methodological and scientific disciplines: epistemology, rationality, ethics, physics, etc. Also, the LLM is given tools: narrow AIs (such as for protein folding or for predicting properties of materials, or for formal scientific modelling). Finally, the LLM is given a compute engine such as Wolfram and a knowledge base such as Wikidata or Wolfram Knowledgebase. The exemplary actor creates plans or predictions for given situations (described in language and fed to the LLM underlying the exemplary actor as prompts) and iteratively critiques and refines its own plans and predictions while putting different textbooks into the LLM context (first, with the textbook on rationality, then epistemology, then physics, etc., with potentially dozens of different textbooks relevant for a plan or prediction that is being criticised), for many iterations, until convergence. In section 2.1, I note that the type of alignment that the exemplary actor's architecture tries to ensure is called (world) model alignment and that is stronger and also more essential than goal alignment. Then, I discuss the properties of the exemplary actor. In section 2.2., I discuss what I see as likely non-issues or straightforwardly addressable issues: the “divergent reasoning nature” of LLMs, the lack of grounded common sense reasoning, and the bias of the quick reactive network (”System 1”), it it is added to the architecture to make it more practically usable in lower-stakes reasoning settings. In section 2.3, I discuss the outstanding technical issues and risks of the exemplary actor's architecture: The risk of direct access to the underlying LLM (section 2.3.1). The exemplary actor's reasoning could still be partially directed by “alien” thinking patterns (i.e., the world model) of the underlying LLM even though these influences won't surface in the explanations of the plan (section 2.3.2). Iterated critique and refinement probably won't make plans strictly conforming to the theories described in the textbooks (section 2.3.3). In section 2.3.4, I discuss the alignment tax of the exemplary actor (compared with the baseline of a bare, minimally fine-tuned LLM) and conclude that the main source of alignment tax might happen to be the theory of ethics which may force the exemplary actor to refuse to participate in “games” (i.e., real-world situations and environments) where it doesn't see ethical ways of “winning”, and thus will consider inaction (or some form of palliative action) the only ethical way forward. This is not a technical problem with the exemplary actor per se, but rather a problem with a higher-level system, i.e., the current economic, social, and political structure of the world. I mention this and other kinds of “higher-level” risks of the plans to build and deploy the exemplary actor (i.e., roughly the plans that OpenAI and Anthropic are betting on, as it seems to me) in section 2.4. 2. An LLM-based “exemplary actor” Let's assume that we have three things: First, a very powerful auto-regressive LLM (think GPT-5/6 level) with the ability to effe...

Funny Science Fiction
S2E111 - A Vorta and a Bajoran Walk into a Spaceport ft. Kitty Swink

Funny Science Fiction

Play Episode Listen Later Dec 9, 2022 49:11


A Vorta and a Bajoran Walk into a Spaceport You may best recognize this week's guest from her work on Babylon 5, Star Trek Deep Space Nine, JAG, NYPD Blue, Becker and more! Kitty Swink is our guest this week, and we are excited to have her join us! Kitty talks about her work in the performing arts over the years, her love for Star Trek and how its impacted her life, her fight with pancreatic cancer and survival along with her work with the Pancreatic Cancer Action Network and more! Be sure to check out her response to WikiData claiming they have proof that she's an actual human! You can find out more about the Pancreatic Cancer Action Network by visiting www.pancan.org For more information about Kitty Swink, please visit the following - Twitter - @kitswink For RSWOF Merch - https://www.teepublic.com/t-shirt/31938193-rswof?store_id=1397534 100% of all proceeds Benefit Wish Upon a Teen For direct contributions - www.wishuponateen.org Join our Discord! https://discord.gg/cpry4fCDTq FSF PopCast on Twitter and Instagram - @fsfpopcast Buy us Coffee - https://ko-fi.com/fsfpopcast For more on our show partners - Big Boy Graphics - www.etsy.com/shop/bigboygraphics Bridgework Studios - https://www.teepublic.com/user/bridgework-studios Level Up Sabers - https://shareasale.com/r.cfm?b=2018189&u=3289465&m=124959&urllink=&afftrack= Support The FSF PopCast by contributing to their tip jar: https://tips.pinecast.com/jar/funny-science-fiction

The FSF PopCast
S2E111 -A Vorta and a Bajoran Walk into a Spaceport ft. Kitty Swink

The FSF PopCast

Play Episode Listen Later Dec 9, 2022 49:11


A Vorta and a Bajoran Walk into a Spaceport You may best recognize this week's guest from her work on Babylon 5, Star Trek Deep Space Nine, JAG, NYPD Blue, Becker and more! Kitty Swink is our guest this week, and we are excited to have her join us! Kitty talks about her work in the performing arts over the years, her love for Star Trek and how its impacted her life, her fight with pancreatic cancer and survival along with her work with the Pancreatic Cancer Action Network and more! Be sure to check out her response to WikiData claiming they have proof that she's an actual human! You can find out more about the Pancreatic Cancer Action Network by visiting www.pancan.org For more information about Kitty Swink, please visit the following - Twitter - @kitswink For RSWOF Merch - https://www.teepublic.com/t-shirt/31938193-rswof?store_id=1397534 100% of all proceeds Benefit Wish Upon a Teen For direct contributions - www.wishuponateen.org Join our Discord! https://discord.gg/cpry4fCDTq FSF PopCast on Twitter and Instagram - @fsfpopcast Buy us Coffee - https://ko-fi.com/fsfpopcast For more on our show partners - Big Boy Graphics - www.etsy.com/shop/bigboygraphics Bridgework Studios - https://www.teepublic.com/user/bridgework-studios Level Up Sabers - https://shareasale.com/r.cfm?b=2018189&u=3289465&m=124959&urllink=&afftrack= Support The FSF PopCast by contributing to their tip jar: https://tips.pinecast.com/jar/funny-science-fiction

Sustain
Episode 147: Jan Ainali and the Foundation for Public Code

Sustain

Play Episode Listen Later Dec 2, 2022 36:42


Guest Jan Ainali Panelists Richard Littauer | Justin Dorfman Show Notes Hello and welcome to Sustain! The podcast where we talk about sustaining open source for the long haul. On this episode, we have an amazing guest, Jan Ainali, who's a steward at the Foundation for Public Code, where they develop tools, processes, and collecting best practices for community building. Earlier, he ran a consultancy called Open by Default and he was swept into the ‘open movement' through Wikipedia editing and leading him to cofound the Swedish Wikimedia chapter and become chairman and later, CEO. Today, we'll learn all the details about the Foundation for Public Code, the process of becoming a member, what sets them apart from others, and their Standard for Public Code that is for anyone who wants to prepare their code to be collaborated on. Also, Jan shares what's he looking forward to in the next few months about the standard and the Foundation for Public Code. Download this episode now to learn much more! [00:02:02] Jan tells us about the Foundation for Public Code, as well as how many member organizations it has. [00:03:43] With only one member currently, we find out if Jan is trying to get others on to work with the foundation, he explains the process of becoming of member with them, and the team sizes that are directly working with them. [00:07:02] We learn what sets apart the Foundation for Public Code from the other trans-provincial and trans-governmental organizations that are doing the work of InnerSource Commons but with politics. He also goes into the policies that have gone into code that he's worked on. [00:09:46] Wikimedia and Wikipedia have chapters, and Jan reveals how big his was, how many other chapters there are in the world, and the difference between them. [00:11:15] Find out who Sverker Johansson is and what he did. [00:13:12] Jan tells us more about the Standard for Public Code, what it is, how it applies, who wrote it, and we hear the 16 criteria for it. [00:18:13] Jan explains the “must be in plain English” requirement and what the global efforts are for the Foundation for Public Code. [00:20:13] We learn how Jan is making it beneficial for everyone to join in a way that helps them out as being super awesome and grow the network that way. [00:21:02] Has Jan gotten any pushback from developers in other places? [00:22:24] Jan tells us businesses he's working with to help push this initiative forward. [00:24:38] We hear if there's a sign on process for getting people to use this standard, and Jan talks about the accreditation process. [00:29:14] Richard asks Jan if he knows of other standards that are already in this space and what sets his apart from the others. [00:30:32] Jan explains their level of standards, as well as what he's most looking forward to in the next few months about the standard and the foundation. [00:32:54] Find out where you can follow Jan on the web. Quotes [00:03:19] “We really think if you collaborate with everyone, that's better than to collaborate with just a few." [00:07:22] “We only work with code bases with the public purpose where someone tries to put a policy into code. That's where we're a little bit narrower than others.” Spotlight [00:34:04] Justin's spotlight is WeasyPrint. It's open source and turns simple HTML pages into gorgeous PDF's and open source. [00:34:39] Richard's spotlight is Sverker Johansson. [00:35:10] Jan's spotlight is Denny Vrandečić, first project manager at Wikidata, and right now working on Abstract Wikipedia. Links SustainOSS (https://sustainoss.org/) SustainOSS Twitter (https://twitter.com/SustainOSS?ref_src=twsrc%5Egoogle%7Ctwcamp%5Eserp%7Ctwgr%5Eauthor) SustainOSS Discourse (https://discourse.sustainoss.org/) podcast@sustainoss.org (mailto:podcast@sustainoss.org) Richard Littauer Twitter (https://twitter.com/richlitt?ref_src=twsrc%5Egoogle%7Ctwcamp%5Eserp%7Ctwgr%5Eauthor) Justin Dorfman Twitter (https://twitter.com/jdorfman?ref_src=twsrc%5Egoogle%7Ctwcamp%5Eserp%7Ctwgr%5Eauthor) Jan Ainali Twitter (https://twitter.com/Jan_Ainali?ref_src=twsrc%5Egoogle%7Ctwcamp%5Eserp%7Ctwgr%5Eauthor) Jan Ainali Website (https://ainali.com/) Foundation for Public Code (https://publiccode.net/) Standard for Public Code (https://standard.publiccode.net/) WeasyPrint (https://weasyprint.org/) Standard for Public Code (Book) (https://github.com/publiccodenet/standard#generating-a-pdf-of-the-standard-for-public-code) Sverker Johansson-Wikipedia (https://en.wikipedia.org/wiki/Sverker_Johansson) [The Dawn of Language by Sverker Johansson](https://www.amazon.com/The-Dawn-of-Language/dp/1529411416/ref=tmmpapswatch0?encoding=UTF8&qid=&sr=) xkcd-Standards (https://xkcd.com/927/) Wikidata (https://www.wikidata.org/wiki/Wikidata:Main_Page) Abstract Wikipedia (https://meta.wikimedia.org/wiki/Abstract_Wikipedia) Credits Produced by Richard Littauer (https://www.burntfen.com/) Edited by Paul M. Bahr at Peachtree Sound (https://www.peachtreesound.com/) Show notes by DeAnn Bahr Peachtree Sound (https://www.peachtreesound.com/) Special Guest: Jan Ainali.

Starke Frauen
#164 Maria Theresia von Österreich - prägende Monarchin der Ära des aufgeklärten Absolutismus

Starke Frauen

Play Episode Listen Later Nov 29, 2022 28:42


Sie war Fürstin aus dem mächtigen Hause Habsburg, das seit dem Mittelalter zahlreiche europäische Herrscher*innen hervor gebracht hat, z.B. alle Kaiser des Heiligen Römischen Reiches (deutscher Nation) bis 1806. Darunter Maria Theresia.Sie war 1740 bis zu ihrem Tod Erzherzogin von Österreich und Königin u.a. von Ungarn (mitKroatien) und Böhmen.Sie war die Gewinnerin des österreichischen Erbfolgekriegs: Bis 1748 konnte sich Maria Theresia im Frieden von Aachen als rechtmäßige Erbin des 1740 verstorbenen Karl VI. durchsetzen, wobei sie trotzdem einige Ländereien wie Schlesien verlor sowie die Herzogtümer Parma und Piacenza.Maria Theresia ist besonders bekannt für politische und gesellschaftliche Reformen, die aus heutiger Sicht sehr modern anmuten, darunter eine größere Bedeutung für Bildung in der Gesellschaft. Damit wurde Maria Theresia eine prägende Figur des aufgeklärten Absolutismus.Mehr zu ihr in dieser Episode.Photo Credits: wiki commons, Deutsch: Maria Theresia von Österreich-Este, Lithographie von Franz Eybl nach einem Gemälde von Moritz Daffinger, 1845Date1845SourceEigenes Foto einer Originallithographie der ÖNB (Wien) Author: Franz Eybl  (1806–1880)  Foto Peter Geymayer at German Wikipedia.Redaktionelle Unterstützung: Daniel Jacob Mehr zu uns unter: https://linktr.ee/starkefrauen****************WERBUNG: Unser Podcast wird freundlicherweise unterstützt von CLARK, dem digitalen Versicherungsmanager. Versichert euch jetzt "up" mit der kostenlosen CLARK-App und erfahrt alle Infos zum Shopping-Gutschein mit dem Code “STARKEFRAUEN“ hier: https://www.clark.de/landing/social/starke-frauen/ Möchtest Du Cathrin oder Kim auf einen Kaffee einladen und dafür die Episoden werbefrei hören? Dann klicke auf den folgenden Link: https://plus.acast.com/s/starke-frauen. Hosted on Acast. See acast.com/privacy for more information.

UVA Data Points
WikiProject Biography

UVA Data Points

Play Episode Listen Later Nov 21, 2022 36:42


This bonus episode features a conversation between Lane Rasberry, Wikimedian-In-Residence at the UVA School of Data Science, and Lloyd Sy, a Ph.D. candidate in the UVA Department of English. In this conversation, Lane and Lloyd take a deep dive into the expansive world of Wikidata and ask the existential question, "What makes a person a person?" Or, more specifically, what data points make up a person? To help answer this question, Lloyd developed a large-scale data model of the biographical data contained within the Wikidata platform. This project serves as the foundation for their conversation. They also take a wide view of biographical data as it pertains to research and academia, including the process of gathering the data, the ethics of utilizing the data, personal ownership of the data, and much more. Anyone interested in these concepts should find this discussion valuable. Links: WikiProject Biography Music: "Screen Saver" Kevin MacLeod (incompetech.com)Licensed under Creative Commons: By Attribution 4.0 Licensehttp://creativecommons.org/licenses/by/4.0/

Nova Ràdio Lloret
Bon Dia Bona Hora – 11 Agost 2022

Nova Ràdio Lloret

Play Episode Listen Later Aug 11, 2022 59:49


Aquest dijous ens fem la següent reflexió: qui és la personalitat més popular nascuda a Lloret de Mar? La personalitat més popular nascuda a Lloret de Mar és el jugador de futbol Marc Muniesa, segons un mapa creat per un geògraf finlandès, a partir de les dades de Viquipèdia i Wikidata. Es tracta d'una informació, […] La notícia Bon Dia Bona Hora – 11 Agost 2022 s'ha publicat al web de Nova Ràdio Lloret.

Infotecarios Podcast
InfoTecarios Podcast 106: Wikipedia, Wikimedia Commons y Wikidata (Con Luis Alvaz)

Infotecarios Podcast

Play Episode Listen Later Jun 20, 2022 71:32


Nueva edición del podcast de www.infotecarios.com En esta edición charlaremos con Luis Álvarez Azcárraga sobre documentación, registro y archivo del espacio público a través de herramientas de conocimiento libre: Wikipedia, Wikimedia Commons y Wikidata. Conocido como Luis Alvaz dentro de la comunidad de Wikimedia México y Wikipedia. Luis es Doctor en Estudios Socioculturales (UAA), maestro en Estética y Arte (BUAP) y licenciado en Ciencias de la Comunicación (BUAP). Profesor en la Licenciatura en Estudios del Arte y Gestión Cultural (UAA) y en la Licenciatura en Artes Visuales (Universidad de las Artes, ICA), desde el 2015. Sus áreas de investigación son: Las comunidades colaborativas y participativas de Internet, como Wikipedia. También ha estudiado los procesos creativos en las artes digitales, así como el arte sonoro y los sistemas de colaboración en el arte y la música. Está interesado en las licencias libres, la cultura copyleft, la ciencia abierta y el conocimiento como un procomún. Desde 2014, edita en Wikipedia y es voluntario de Wikimedia México desde ese mismo año, organizando talleres, editatones e incorporando esta herramienta como elemento para el aprendizaje en la universidad. Forma parte de la mesa directiva de Wikimedia México desde el año 2016, donde es vicepresidente desde este año. También es integrante de Creative Commons México desde el año 2020. Sus actividades artísticas: Ha musicalizado documentales y obras de teatro, como Corazón gordito (Teatro escolar Aguascalientes, 2016-2017) y Clarita (Muestra Nacional de Teatro 2019), con la colectiva Ramas y Raíces. Colaborador del podcast Se Oye Libre de Wikimedia México (2021-2022). Productor, editor y locutor de radio: El Grito del Silencio (2010-2011); El Ágora (RadioBuap, 2008); Comunicación al Aire (RadioBuap, 2006-2007) Es editor de sección de la Revista Arte Imagen y Sonido, de la Universidad Autónoma de Aguascalientes.

Podcast Libre à vous !
Software Heritage

Podcast Libre à vous !

Play Episode Listen Later Mar 1, 2022 58:10


Les références : Software Heritage Site de Morgane Gruenpeter Software Stories (utilise Wikidata) Save code now Les ambassadeurs de Software Heritage Postuler à Software Heritage Guide de bonnes pratiques Software Heritage archive Internet archive Libre à vous ! #13 du 19 février 2019, avec Roberto Di Cosmo, un des fondateurs du projet Software Heritage. Libre à vous ! #130 sur la science ouverte et logiciel libreVous pouvez commenter les émissions, nous faire des retours pour nous améliorer, ou encore des suggestions. Et même mettre une note sur 5 étoiles si vous le souhaitez. Il est important pour nous d'avoir vos retours car, contrairement par exemple à une conférence, nous n'avons pas un public en face de nous qui peut réagir. Pour cela, rendez-vous sur la page dédiée.

Her Royal Science
26 Community

Her Royal Science

Play Episode Listen Later Jan 26, 2022 27:52


In this episode, we speak with Dr Sabah Ul-Hasan, a bioinformatics postdoctoral scholar and lecturer who curates and integrates specialised medical databases into Wikidata. They are also one of the founders of The Biota Project, which intersects science education, outreach, and environmental justice to foster science and data literacy. We begin our conversation by exploring their first moments of fascination with science, and in particular, with wildlife biology. Later on, we discuss the value of data accessibility, the importance of community, and the valuable lessons that Dr Ul-Hasan has learned along their journey. Episode transcript available here: https://www.herroyalscience.com/post/26-community

Conversas em Código
Ep. 40 - Nada funciona nunca

Conversas em Código

Play Episode Listen Later Jan 20, 2022 17:57


Uma semana cheia de problemas, outra vez no Percy, nas contribuições do Ember, e mais umas esquisiticesSegue-nos no Mastodon ou no Twitter e junta-te ao nosso Slack.Links:WikidataEmber.js na wikidataGithub wikidata bottool-new-releaseFirefox multi-account containersPercyChoosyBufferMastodon schedulerO Conversas em Código é da autoria do Hugo Peixoto e de Ricardo Mendes

From where does it STEM?
Objectivity vs. Subjectivity in Science: Dr. Sabah Ul-Hasan

From where does it STEM?

Play Episode Listen Later Nov 9, 2021 43:46


On this episode, I got the chance to sit down and chat about social justice in science with Dr. Sabah Ul-Hasan, a bioinformatician and postdoc at The Scripps Research Institute in San Diego under Dr. Andrew Su. They are currently working on the NIH-funded Wikidata biocuration project. Enjoy! --- Send in a voice message: https://podcasters.spotify.com/pod/show/fromwheredoesitstem/message

The World According to Wikipedia
S2 Ep20: That Wikidata Buzz

The World According to Wikipedia

Play Episode Listen Later Sep 22, 2021 34:21


In this episode, we will talk to Siobhan Leachman, a citizen scientist and prolific Wikidata editor. Rebecca tries to figure out which WikiFauna represents her best. And our Wiki hero is the incoming CEO Maryana Iskander. Logo design by Trish O'Flaherty: https://www.trishoflahertydesign.com/ Twitter: https://twitter.com/world_wikipedia Website: https://www.headstuff.org/the-world-according-to-wikipedia/ This show is part of the HeadStuff Podcast Network. For more, go to HeadStuffPodcasts.com, where you can also become a member of HeadStuff+ and get exclusive access to bonus material and lots more.

The World According to Wikipedia
S2 Ep19: Stanning Wikidata

The World According to Wikipedia

Play Episode Listen Later Sep 8, 2021 33:41


In this episode, we will talk to Jan Ainali, a Codebase Steward for the Foundation for Public Code and a complete Wikidata fanboy. The wiki-rule of the episode is that no-one owns an article and we celebrate the many Wikimedians of 2021, including past guest Netha Hussain.  Some interesting links from the Interview with Jan: https://www.datastory.org/services/election-tracker https://www.govdirectory.org/ Logo design by Trish O'Flaherty: https://www.trishoflahertydesign.com/ Twitter: https://twitter.com/world_wikipedia Website: https://www.headstuff.org/the-world-according-to-wikipedia/ This show is part of the HeadStuff Podcast Network. For more, go to HeadStuffPodcasts.com, where you can also become a member of HeadStuff+ and get exclusive access to bonus material and lots more.

R Weekly Highlights
Issue 2021-W25 Highlights

R Weekly Highlights

Play Episode Listen Later Jun 23, 2021 12:03


Projecting and tracking COVID-19 infection rates in England with R, leveraging Wikidata to tag scientific abstracts, and a new deep-learning workflow with the luz package Episode Links This week's curator: Robert Hickman (@robwhickman (https://twitter.com/robwhickman)) Tracking SARS-CoV-2 In England with {epidemia} (https://imperialcollegelondon.github.io/epidemia/articles/multiple-obs.html) Tagging the Scientific Abstracts with Wikidata Items (https://dwayzer.netlify.app/posts/2021-06-15-tagging-the-abstracts-with-wikidata-items) Que haja luz: More light for torch! (https://blogs.rstudio.com/tensorflow/posts/2021-06-17-luz) Entire issue available at rweekly.org/2021-W25 (https://rweekly.org/2021-W25.html) Supplemental Resources {epidemia} package documentation (https://imperialcollegelondon.github.io/epidemia/index.html) A COVID-19 Model for Local Authorities of the United Kingdom (https://rss.org.uk/RSS/media/File-library/News/2021/MishraScott.pdf) How epidemiology has shaped the COVID pandemic (https://www.nature.com/articles/d41586-021-00183-z)

R Weekly Highlights
Issue 2021-W17 Highlights

R Weekly Highlights

Play Episode Listen Later Apr 27, 2021 12:37


Exploring wikidata with the {tidywikidatar} package, accessibility improvements in {knitr}, and the top 40 new CRAN packages for March. Episode Links This week's curator: Batool Almazrouq (@batool664) (https://twitter.com/batool664) What does Wikidata know about members of the European Parliament? (https://medium.com/european-data-journalism-network/a-new-r-package-for-exploring-the-wealth-of-information-stored-by-wikidata-fe85e82b6440) New in knitr: Improved accessibility with image alt text (https://blog.rstudio.com/2021/04/20/knitr-fig-alt/) March 2021: "Top 40" New CRAN Packages (https://rviews.rstudio.com/2021/04/22/march-2021-top-40-new-cran-packages/) Entire issue available at rweekly.org/2021-W17 (https://rweekly.org/2021-W17.html) Supplemental Resources https://news.yahoo.com/wikipedia-turns-20-aims-reach-035212015.html https://github.com/yihui/knitr/releases/tag/v1.32 https://cran.r-project.org/web/packages/pkglite/index.html https://flujoo.github.io/gm/

With Jason Barnard...
Advantages of Structured Data (Jarno van Driel and Jason Barnard)

With Jason Barnard...

Play Episode Listen Later Apr 6, 2021 43:56


Jarno van Driel talks with Jason Barnard about structured data Jarno Van Driel is a technical & semantic SEO expert with more than 20 years of experience in the field, whose main focus is to provide website optimization services and guide organizations into this new era of the semantic web. Contrary to what most marketers believe, structured data is much more than a simple final touch to get a nice Brand SERP – there are serious advantages in making structured data the foundation for your organization. You will learn amazing tips and insights, as well as what the future looks like in terms of the implementation of Schema Markup in e-commerce sites. What you'll learn from Jarno van Driel 00:00 Jarno Van Driel with Jason Barnard00:40 Jarno Van Driel's Brand SERP02:35 Structured Data is MUCH more than just a “final sprinkle” to get a nice looking SERP07:02 Structured Data as the foundation for your organization08:36 Managing Data Layers in Google Tag Manager11:15 How to deal with the current Schema Markup limitations?13:58 Could Structured Data solve the language barrier issue on a search engine?15:54 Wikipedia, Wikidata and Dbpedia for entity disambiguation in different languages17:15 Is there always a need to disambiguate?17:53 What has changed in SEO copywriting since 201020:26 Does Google prefer that a page is about just one single entity (passage based indexing) ?22:33 A discussion about Google's ability to identify content's relevancy27:54 Using Schema Markup to avoid duplicate content on an e-commerce site at the category level32:22 New implementations of Schema Markup in e-commerce sites38:09 Google's Merchant Center feeds and the role of Schema Markup41:06 Is Google “forcing our hand” with structured data? Subscribe to the podcast Subscribe here >> This episode was recorded live on video April 6th 2021 Recorded live at Kalicube Tuesdays (Digital Marketing Livestream Event Series). Watch the video now >>

With Jason Barnard...
Advantages of Structured Data (Jarno van Driel and Jason Barnard)

With Jason Barnard...

Play Episode Listen Later Apr 6, 2021 43:56


Jarno van Driel talks with Jason Barnard about structured data Jarno Van Driel is a technical & semantic SEO expert with more than 20 years of experience in the field, whose main focus is to provide website optimization services and guide organizations into this new era of the semantic web. Contrary to what most marketers believe, structured data is much more than a simple final touch to get a nice Brand SERP – there are serious advantages in making structured data the foundation for your organization. You will learn amazing tips and insights, as well as what the future looks like in terms of the implementation of Schema Markup in e-commerce sites. What you'll learn from Jarno van Driel 00:00 Jarno Van Driel with Jason Barnard00:40 Jarno Van Driel's Brand SERP02:35 Structured Data is MUCH more than just a “final sprinkle” to get a nice looking SERP07:02 Structured Data as the foundation for your organization08:36 Managing Data Layers in Google Tag Manager11:15 How to deal with the current Schema Markup limitations?13:58 Could Structured Data solve the language barrier issue on a search engine?15:54 Wikipedia, Wikidata and Dbpedia for entity disambiguation in different languages17:15 Is there always a need to disambiguate?17:53 What has changed in SEO copywriting since 201020:26 Does Google prefer that a page is about just one single entity (passage based indexing) ?22:33 A discussion about Google's ability to identify content's relevancy27:54 Using Schema Markup to avoid duplicate content on an e-commerce site at the category level32:22 New implementations of Schema Markup in e-commerce sites38:09 Google's Merchant Center feeds and the role of Schema Markup41:06 Is Google “forcing our hand” with structured data? Subscribe to the podcast Subscribe here >> This episode was recorded live on video April 6th 2021 Recorded live at Kalicube Tuesdays (Digital Marketing Livestream Event Series). Watch the video now >>

República Web
Wikimedia es mucho más que Wikipedia con Rubén Ojeda

República Web

Play Episode Listen Later Oct 25, 2020 78:00


Para este episodio contamos con la compañía de Rubén Ojeda, coordinador de proyectos en Wikimedia España, la asociación sin ánimo de lucro que promueve el conocimiento libre y los proyectos Wikimedia, siendo Wikipedia el más conocido de todos. Pero Wikimedia es mucho más que la Wikipedia, y en este episodio Rubén nos contará todo el trabajo que se realiza desde la asociación y cómo favorecen un mejor acceso a la cultura y el conocimiento libre. Wikimedia está detrás de proyectos de difusión tan interesantes como Wikidata, Wikimedia Commons, Wikiquote o Wikisource, pero también de otros muchos que constituyen un formidable repositorio de información compartida y libre. Con Rubén Ojeda hablamos mucho sobre Wikipedia pero también de varias cuestiones sobre Wikimedia y sus proyectos: ¿Qué es la Asociación Wikimedia España y cómo se constituye aquí? En qué consiste una wiki y cómo se organiza a nivel editorial. ¿Cómo pueden participar los ciudadanos de los proyectos Wikimedia? ¿Conoce el sector público y las instituciones culturales cómo compartir conocimiento a través de vuestros proyectos? ¿Cuáles son las mayores amenazas al conocimiento libre en la era digital? ¿Cómo pueden ayudar los proyectos Wikimedia a combatir la desinformación? En definitiva una conversación muy completa para conocer la labor de Wikimedia, aspectos sobre derechos digitales, contenidos libres y conocer cómo poder colaborar en los proyectos.

The History of Computing

Welcome to the History of Computing Podcast, where we explore the history of information technology. Because understanding the past prepares us for the innovations of the future! Todays episode is on the history of Wikipedia. The very idea of a single location that could store all the known information in the world began with Ptolemy I, founder of the Greek dynasty that ruled Egypt following the death of Alexander the great. He and his son amassed 100s of thousands of scrolls in the Library and Alexandria from 331 BC and on. The Library was part of a great campus of the Musaeum where they also supported great minds starting with Ptolemy I's patronage of Euclid, the father of geometry, and later including Archimedes, the father of engineering, Hipparchus, the founder of trigonometry, Her, the father of math, and Herophilus, who gave us the scientific method and countless other great hellenistic thinkers. The Library entered into a slow decline that began with the expulsion of intellectuals from Alexandria in 145BC. Ptolemy VIII was responsible for that. Always be weary of people who attack those that they can't win over especially when they start blaming the intellectual elite for the problems of the world. This began a slow decline of the library until it burned, first with a small fire accidentally set by Caesar in 48BC and then for good in the 270s AD. In the centuries since there have been attempts here and there to gather great amounts of information. The first known encyclopedia was the Naturalis Historiae by Pliny the Elder, never completed because he was killed in the eruption of Vesuvius. One of the better known being the Encyclopedia Britannica, starting off in 1768. Mass production of these was aided by the printing press but given that there's a cost to producing those materials and a margin to be made in the sale of those materials that encouraged a somewhat succinct exploration of certain topics. The advent of the computer era of course led to encyclopedias on CD and then to online encyclopedias. Encyclopedias at the time employed experts in certain fields and paid them for compiling and editing articles for volumes that would then be sold. As we say these days, this was a business model just waiting to be disrupted. Jimmy Wales was moderating an online discussion board on Objectivism and happened across Larry Sanger in the early 90s. They debated and became friends. Wales started Nupedia, which was supposed to be a free encyclopedia, funded by advertising revenue. As it was to be free, they were to recruit thousands of volunteer editors. People of the caliber that had been previously hired to research and write articles for encyclopedias. Sanger, who was pursuing a PhD in philosophy from Ohio State University, was hired on as editor-in-chief. This was a twist on the old model of compiling an encyclopedia and a twist that didn't work out as intended. Volunteers were slow to sign up, but Nupedia went online in 2000. Later in the year there had only been two articles that made it through the review process. When Sanger told Ben Kovitz about this, he recommended looking at the emerging wiki culture. This had been started with WikiWikiWeb, developed by Ward Cunningham in 1994, named after a shuttle bus that ran between airport terminals at the Honolulu airport. WikiWikiWeb had been inspired by Hypercard but needed to be multi-user so people could collaborate on web pages, quickly producing content on new patterns in programming. He wanted to make non-writers feel ok about writing. Sanger proposed using a wiki to be able to accept submissions for articles and edits from anyone but still having a complicated review process to accept changes. The reviewers weren't into that, so they started a side project they called Wikipedia in 2001 with a user-generated model for content, or article, generation. The plan was to generate articles on Wikipedia and then move or copy them into Nupedia once they were ready. But Wikipedia got mentioned on Slashdot. In 2001 there were nearly 30 million websites but half a billion people using the web. Back then a mention on the influential Slashdot could make a site. And it certainly helped. They grew and more and more people started to contribute. They hit 1,000 articles in March of 2001 and that increased by 10 fold by September, By And another 4 fold the next year. It started working independent of Nupedia. The dot-com bubble burst in 2000 and by 2002 Nupedia had to lay Sanger off and he left both projects. Nupedia slowly died and was finally shut down in 2003. Eventually the Wikimedia Foundation was built to help unlock the world's knowledge, which now owns and operates Wikipedia. Wikimedia also includes Commons for media, Wikibooks that includes free textbooks and manuals, Wikiquote for quotations, Wikiversity for free learning materials, MediaWiki the source code for the site, Wikidata for pulling large amounts of data from Wikimedia properties using APIs, Wikisource, a library of free content, Wikivoyage, a free travel guide, Wikinews, free news, Wikispecies, a directory containing over 687,000 species. Many of the properties have very specific ways of organizing data, making it easier to work with en masse. The properties have grown because people like to be helpful and Wales allowed self-governance of articles. To this day he rarely gets involved in the day-to-day affairs of the wikipedia site, other than the occasional puppy dog looks in banners asking for donations. You should donate. He does have 8 principles the site is run by: 1. Wikipedia's success to date is entirely a function of our open community. 2. Newcomers are always to be welcomed. 3. “You can edit this page right now” is a core guiding check on everything that we do. 4. Any changes to the software must be gradual and reversible. 5. The open and viral nature of the GNU Free Documentation License and the Create Commons Attribution/Share-Alike License is fundamental to the long-term success of the site. 6. Wikipedia is an encyclopedia. 7. Anyone with a complaint should be treated with the utmost respect and dignity. 8. Diplomacy consists of combining honesty and politeness. This culminates in 5 pillars wikipedia is built on: 1. Wikipedia is an encyclopedia. 2. Wikipedia is written from a neutral point of view. 3. Wikipedia is free content that anyone can use, edit, and distribute. 4. Wikipedia's editors should treat each other with respect and civility. 5. Wikipedia has no firm rules. Sanger went on to found Citizendium, which uses real names instead of handles, thinking maybe people will contribute better content if their name is attached to something. The web is global. Throughout history there have been encyclopedias produced around the world, with the Four Great Books of Song coming out of 11th century China, the Encyclopedia of the Brethren of Purity coming out of 10th century Persia. When Wikipedia launched, it was in English. Wikipedia launched a German version using the deutsche.wikipedia.com subdomain. It now lives at de.wikipedia.com and Wikipedia has gone from being 90% English to being almost 90 % non-English, meaning that Wikipedia is able to pull in even more of the world's knowledge. Wikipedia picked up nearly 20,000 English articles in 2001, over 75,000 new articles in 2002, and that number has steadily climbed wreaching over 3,000,000 by 2010, and we're closing in on 6 Million today. The English version is 10 terabytes of data uncompressed. If you wanted to buy a printed copy of wikipedia today, it would be over 2500 books. By 2009 Microsoft Encarta shut down. By 2010 Encyclopedia Britannica stopped printing their massive set of books and went online. You can still buy encyclopedias from specialty makers, such as the World Book. Ironically, Encyclopedia Britannica does now put real names of people on articles they produce on their website, in an ad-driven model. There are a lot of ads. And the content isn't linked to as many places nor as thorough. Creating a single location that could store all the known information in the world seems like a pretty daunting task. Compiling the non-copywritten works of the world is now the mission of Wikipedia. The site receives the fifth most views per month and is read by nearly half a billion people a month with over 15 billion page views per month. Anyone who has gone down the rabbit hole of learning about Ptolemy I's involvement in developing the Library of Alexandria and then read up on his children and how his dynasty lasted until Cleopatra and how… well, you get the point… can understand how they get so much traffic. Today there are over 48,000,000 articles and over 37,000,000 registered users who have contributed articles meaning if we set 160 Great Libraries of Alexandria side-by-side we would have about the same amount of information Wikipedia has amassed. And it's done so because of the contributions of so many dedicated people. People who spend hours researching and building pages, undergoing the need to provide references to cite the data in the articles (btw wikipedia is not supposed to represent original research), more people to patrol and look for content contributed by people on a soapbox or with an agenda, rather than just reporting the facts. Another team looking for articles that need more information. And they do these things for free. While you can occasionally see frustrations from contributors, it is truly one of the best things humanity has done. This allows us to rediscover our own history, effectively compiling all the facts that make up the world we live in, often linked to the opinions that shape them in the reference materials, which include the over 200 million works housed at the US Library of Congress, and over 25 million books scanned into Google Books (out of about 130 million). As with the Great Library of Alexandria, we do have to keep those who seek to throw out the intellectuals of the world away and keep the great works being compiled from falling to waste due to inactivity. Wikipedia keeps a history of pages, to avoid revisionist history. The servers need to be maintained, but the database can be downloaded and is routinely downloaded by plenty of people. I think the idea of providing an encyclopedia for free that was sponsored by ads was sound. Pivoting the business model to make it open was revolutionary. With the availability of the data for machine learning and the ability to enrich it with other sources like genealogical research, actual books, maps, scientific data, and anything else you can manage, I suspect we'll see contributions we haven't even begun to think about! And thanks to all of this, we now have a real compendium of the worlds knowledge, getting more and more accurate and holistic by the day. Thank you to everyone involved, from Jimbo and Larry, to the moderators, to the staff, and of course to the millions of people who contribute pages about all the history that makes up the world as we know it today. And thanks to you for listening to yet another episode of the History of Computing Podcast. We're lucky to have you. Have a great day! Note: This work was produced in large part due to the compilation of historical facts available at https://en.wikipedia.org/wiki/History_of_Wikipedia

WIRED Business – Spoken Edition
Inside the Alexa-Friendly World of Wikidata

WIRED Business – Spoken Edition

Play Episode Listen Later Feb 19, 2019 7:30


Humans pricked by info-hunger pangs used to hunt and peck for scraps of trivia on the savanna of the internet. Now we sit in screen-glow-flooded caves and grunt, “Alexa!” Virtual assistants do the dirty work for us. Problem is, computers can't really speak the language. Many of our densest, most reliable troves of knowledge, from Wikipedia to (ahem) the pages of WIRED, are encoded in an ancient technology largely opaque to machines—prose.

The Quiet Light Podcast
How to Uncover SEO Opportunities in an Acquisition

The Quiet Light Podcast

Play Episode Listen Later Jun 6, 2018 37:11


I had a buyer recently tell me: "I don't trust SEO traffic. You can't change it. I only trust paid acquisition channels." Times certainly have changed since just a few years ago. A lot of buyers look at natural SEO traffic as untrustworthy ever since the major index updates of Panda, penguin, and hummingbird. Others see SEO as a much more difficult (and possibly more expensive) avenue towards traffic. And some buyers think that the relevance of SEO will be discounted with new voice-enabled searches and paid advertising pushing natural search rankings down Google's SERPs. In this conversation with Corey Northcutt, we discuss the future of SEO, whether there is a good opportunity for buyers to exploit SEO opportunities, and what he would look out for on the SEO front before buying any online business. Episode Highlights Can we trust SEO long-term? Where is it going, and can you build a business on SEO traffic? What key SEO factors should you look at before buying an online business Google's primary goal, and why it is good for business owners Why SEO traffic will always beat paid traffic Why the major Google updates (Panda, Penguin, etc) were a good thing (and still are a good thing) Will paid ads continue to push organic rankings out of SERPs? Are voice-based search devices going to destroy SEO? Click through rates for top rankings in organic listings vs. paid listings Transcription Joe: Hey Mark, how are you doing today? Mark: Doing great. Joe: I understand you recorded a podcast with Corey Northcutt and it was all about SEO. Mark: That's right so Corey and I know each other through Young Entrepreneurs Council which is a great collection of entrepreneurs, some of the best resources that you can find online for anybody that is an entrepreneur out there and is looking for good networking opportunities. You do have to meet certain thresholds in order to join and Corey obviously hits those. I've used his services for another business of mine and was really really impressed with what he had to offer. Brad who works with us is actually the one who first recommended him to me. This guy has been working in the online world and SEO capacity forever. I mean he's a dinosaur in the SEO world, knows a ton about SEO. And one of the questions I posed for him and I really wanted to drill down on this podcast was whether or not there is value at all in SEO anymore from a buy in standpoint; in other words can it be trusted? I had a buyer tell me just a few weeks ago we were talking on the phone about one of the businesses I was representing and he wanted to know where our customers are coming from on the websites and I said well he's got good rankings and then he also does paid. And he came back and he told me and said I don't really care about the organic rankings because I can't control that at all. In fact all I care about is a paid acquisition channel. I think we hear that more and more from people that they trust that paid acquisition channel more than the organic channel. But I think it's almost an overreaction and there's a lot of opportunity being lost because everyone focuses just on the paid and just use the organic just has a sort of bonus to everything else. The couple of the topics that we addressed in this podcast is whether or not the future of SEO is going to be strong. You know we have more and more devices being added for voice-based search. We have the paid creep that's been happening on organic rankings where organic listings are being pushed down the page. And so this is really kind of a step back to look at your online business and say should I be focusing on SEO, what's the opportunity here with SEO as well. Joe: Okay so I'm not going to say whether I think you know you should be focused on it or not. What I can say is that a business with multiple traffics of multiple channels of revenue is worth a lot more money and from my experience a long time ago SEO was a long term game. I survived the Penguin updates, Panda update … actually I sold before the Penguin Update but it didn't matter because I didn't know anything about link building anyway. All I did was good quality content over a long period of time and I was rewarded. It's a lot more complicated than that I think. So I'm really excited to see what Corey has to say. Let's go ahead listen. Mark: Well one second, I'm going to give away a bit of a teaser on this and that is something that you said; I think what you said is perfect because it is the sense that a lot- Joe: Did you say that what I said was perfect just now? Mark: Dude, don't let it get to your head. Yeah it was actually perfect because a lot of people think that SEO is more complex. They think it's complicated. They think that it's a nut that's very difficult to crack. And a lot of us, especially those who have been around through the panda, penguin, hummingbird, you know all these sort of updates; animal updates that happened look at a SEO and like my goodness you can't trust this. Look what happened back then it got completely destroyed. One of the big takeaways from this and Corey gets into this is that SEO now is more predictable than it's ever been. Google's gotten better at what they're doing so they don't need to shake things up as much anymore. Now is a little bit harder to rank; sure because Google has done a better job of putting good information in front of people. But rankings are more stable now than they've ever been before. And so you need big takeaway is that while it's somewhat complex there's a huge opportunity because people like you and I have kind of looked at SEO as that thing as a bonus out there. Joe: From a buyer's point of view what you just said might be very valuable. You know it takes … it's harder to rank now and so if we're listening to business and there's good organic rankings, that in itself could have more value to a buyer. Because most everybody else just cheats to get to the top and you can't actually do that anymore but if you're at the top on page one that's really strong strong value. And I think hopefully in the long run for any business will add more diverse revenue channels which brings its a valuable. Mark: All right can we listen to Corey now? Joe: Absolutely, let's go. Mark: All right Corey, thank you for joining me on this quick episode here on SEO. Corey: Absolutely, thanks for having me. Mark: So yeah I know you've listened to a couple of the episodes before so you know that we like to have our guests introduce themselves. So why don't we provide everyone listening here with a background on yourself. Corey: Yeah I'd love to. So I've been doing SEO for going on 17 years. I've been … I guess running business of different shapes and forms for about as long. I came from web hosting and doing different IT brands and now I run an SEO agency called Northcutt. Mark: That's awesome. Yeah and I've actually … just full disclosure I have user services in the past but with a lot of the guests that we have here on the Quiet Light Podcast, the people that we've use in one cast in another their services so we trust the services, we trust you as far as your … the quality work that you do. And I also know Brad who works with Quiet Light Brokerage; he has used your services quite a bit in the past as well which is how I was referred to you. It was actually through Brad. So that that's pretty cool. Now you have a past in we posting as well is that right? Corey: That's true. Yeah I started at a provider called Ubiquity that was eventually acquired by LeaseWeb but not before I actually exited the company and sold it back to my business partners. Mark: That's pretty cool, so my background is actually web hosting as well. My first job out of college was with a company called Alabanza Corporation and they were the first ones to create the people know like web hosting manager or cPanel. They were the first ones to actually come out with a cPanel and cPanel's a competition to Alabanza. And I remember when we developed that the CEO talking about competition told me he's like I'm not worried about competition, it take … took us years to be able to create our I think he called it the account management [inaudible 00:07:01.2] stupid like that. And of course the temple was already out there so it didn't take them years to replicate what we were doing and it done quite well; so cautionary tale there. But the first business I sold under Quiet Light Brokerage was a web hosting company, so very familiar with the space. So you sold that, how many businesses have you bought and sold over the years? Corey: So I like to call myself a three time founder in all my bios. It's tough to say though because I have partners and in that I've had a lot of failed ventures too. A lot of projects that I've spun up or that maybe sold for cheap so it's been all over the map. I did web hosting. We spun off a data center services brand from that. We had Ventrilo provider called DarkStar Communications which was the largest provider of Ventrilo for quite a while. Most people I think don't even know what that is anymore. Mark: Yeah I've never heard of it. Corey: It was a big bank for a while. And it wasn't even my world but it was kids play World of Warcraft, they would need voice chat for that and it sounds insane today we've got Zoom and Skype and all these tools that are free but they would pay for it. So you'd have 50 or 100 people on one voice server and you needed tools to manage it. So a lot of different businesses; I kind of created a framework for how I like to build and market them but all in completely different spaces. Mark: You know I like to segment Internet entrepreneurs as to those that were started before the panda, penguin days and those that have come into it after because the world is so much different. You know people like you and I that have been in the online world for I'm going on 20 years here since I've built my first website. And you know back then and actually back when I started Quiet Light Brokerage, when I decided to start Quiet Light Brokerage what did I do? I went out and I built a website from scratch and then I custom coded an affiliate program in there and that was kind of how I launched everything was me going out coding, designing, launching, doing the SEO; everything top to bottom. And today I mean if you want to start a new site you can still do that but boy it's so much different today you don't … you really wouldn't want to take that approach as much. It's gotten a lot more complex. Corey: It's true. Mark: Yeah. And you seem to come from that sort of past entrepreneur pay and this is a good idea let me see if I can just build this out real quick. Corey: Yeah and yet it's changed so much. Yeah and I have … on one hand I love how quickly you can spend things up like you've got Shopify, you see people with stores in 10 minutes; it's completely insane. On the other hand I feel like a lot of people don't go as deep with their businesses now. You've got projects I can start a company and it might make 500 dollars a month and that's fine and it … I'll do 10 of those this week. The mindset has changed. Mark: Yeah I know definitely it's changed quite a bit. So I wanted to have you on to talk about SEO and I'll tell you kind of the question that spurred this on for me and I think it would be a good start for our discussion here. I was talking to a buyer the other day about one of my clients and he's Amazon and Magento mixed so he's got his own websites but you know a good portion of his business comes through Amazon. But we're looking at the websites because they're doing really well. Amazon is struggling but the websites are good; really really well. And we're looking at the host and this buyer was asking where does the client … where did the clients come from, are they coming from organic rankings or they're coming from paid service or good mature search campaign out there. So I told him well it's a mix you know there's really good SEO on the site, there's room to improve that as well but they also have a paid campaign that they will get. And this buyer almost seemed to dismiss the SEO side and said well I can't control the SEO world at all, I'm interested in the paid acquisition. And I see this more and more and I think this is kind of the people waking up from the hangover of the panda, penguin which is almost quick as far of be coming up on seven years or something like that; is that right? Seven years is that pretty close? Corey: Well- Mark: [crosstalk 00:11:06.9] try to also work. It's been a while since this happened and I think people have really adjusted their mindset to not trusting SEO at all. So my question to you and to ask in behalf of everyone that out there looking to buy an online business, can you trust SEO and can you build or grow a business on the back of SEO and have it be sustainable? Corey: Oh my God Yes. So there is one question that has been around since I feel like SEO began which is where this is going, it … can I trust it long term, and I think Google's actually been very transparent about what they want to accomplish. As much as we've had different updates like Panda, and Penguin, Hummingbird, chip things up to there's been quite a few but Google's always been forthcoming. And I feel like most of the media out there sensationalizes what's happening and it does a disservice to business owners. Because … and what does Google really care about? They want to reward an experience that is naturally relevant, popular, and enjoyable; that's it. And they've been working towards this goal for all of this time and I don't think it goes away. There is never going to be a point to where a better experience is paid advertising for what they deliver. If they ever reach that point I think all bets are off. I think somebody disrupts them; being or somebody else overtakes them. There's no way people want that. It is a better experience when it's not simply rented. So that doesn't go away it's just they keep it iterating towards getting better at what they set out to do. Like we talk about that pre-panda, pre-penguin world, I think it did a lot of good. I look back at how I did SEO back then and you know it was a little gray. It was hard to … like that was the conversation we were having with people. It's like I think we should be as white hat as possible. At that time I feel like that that was a source of a lot of that grand fish can spam. At the time he was not getting very much respect from professional SEO's and he was saying no, completely white hat, don't mess with anything, no schemes, no link real pyramid tetrahedron. Yeah like I'm sure you've seen all the different diagrams and wacky ideas that people were coming up with back then and that panda made him correct that just overnight everything shifted and it was like well yeah they finally got better and they really are rewarding people that aren't going against Google like do you want to work with them. Mark: Let me play devil's advocate a little bit here and argue against SEO. Now this is not my personal position. I actually agree with you. I think there's a ton of opportunity in SEO and I actually think the world is a … the SEO world is a lot more stable today than it was back in the pre updates of pre-panda, penguin, hummingbird updates mainly because the results are better and Google is still having a better experience and before it was very easy to came, the search engines. You were doing grey hat, I was doing grey hat, everybody was doing grey hat back then. But anyways let me play devil's advocate. Two changes that people look at and they see it as encroaching on the organic SEO. One would be the number of paid listings that show up above the fold on Google and where organic rankings start to be pushed down. And two voice enabled search. Let's start with the first one here the placement of organic rankings. I have another business that I own and I absolutely absolutely hate bidding on my brand keyword. Because it's my brand keyword, I show up number one, I show up number two, I show up number three, but if I don't bid on it I I've got four other people bidding on my brand and they're above me. So from that standpoint has SEO become less valuable for business owners or is that a trend that you think is going to continue where paid listings push out the organic rankings? Corey: I don't and in fact I … you know I saw the same trends. And by the way if somebody just for the benefit of listeners, if somebody is pushing you out that way on your branded searches if they mention you by name you can file a trademark request with Google and get them shut down. It … they're still able to use your brand name as a keyword but it can water down their messaging if somebody is getting too aggressive with that. So I don't see it going too much further and yeah that was a big story each time Google has experimented with expanding the ad block but there's data on quick relates that's out there. Rand Fishkin actually threw his new startup SparkToro all those Jumpshot's analysis on this and it's incredible how many people still click on organic overpaid. The overwhelming majority still click on organic across the board even in the most extreme like biased examples I've seen. I actually just sent out our quarterly here a few weeks ago that looked at this AAReps had their own click through rate data of tons of searches that they've scraped and in their example they said the maximum went up to 46% with click on ads. Up to is the operative word there I think that's the most extreme example. It's 46% where you know it's a branded, your brand is number one; obviously, that's what they wanted. We're going to click that sure but the Jumpshot data said 3% was their average. So somewhere between 3% and 46% are clicking on paid ads. It's still the minority and I don't see that ever changing. Mark: 3% to 46% is pretty broad. Corey: It is pretty broad. The average is three. Mark: Okay. Corey: But yeah. Mark: That's amazing. Yeah and in the example I have we have a lot of brand confusion in our space and my main competitor has been very very good at causing brand confusion. So it's a personal annoyance for me right now, my personal mission to get them out of that number one spot even though I'm losing money on it. Corey: Not this. Mark: Yeah. So that's interesting. I would have tend to agree that there's only so much real estate that they're going to give to the ad spot, to those ad blocks because it's … they have been focused from day one on that user experience. So they want users especially brand searches to be able to find the brand that they're looking for. What about voice enabled search? I know for myself and if I want to find out some quick data or whatever I've got a Google Pixel Phone I just give a little squeeze and Google's system comes up and I just ask it the question and more and more it's becoming intelligent in giving the response. More and more it's taking those responses of course from other websites and so they aren't getting any of the traffic to that. This wouldn't be so much a concern for e-commerce sites but for content sites I mean is this something to be looking out for and maybe something that's going to encroach on their opportunities in the future? Corey: Yeah there is definitely demographics that are going to be hit by this. You and I talked about famous quotations here a week ago in how that is an industry that got hit pretty hard by Panda. I think the nuance is any short simple information is going to have a hard time. Like just the example from last month, Google actually started returning no results searches and people asked for the time of day. And there was a website that was timeanddate.com it ranked number one for all of these and I'm sure they were raking in a lot of AdSense doing it, not probably great ads for those people but still it was working for them in the moment. So there are really nuanced types of businesses that I think buyers should probably be a little wary of. If it doesn't give deep information it can't be [inaudible 00:19:11.5] by a simple answer from Google. But if it does go deeper I think it goes outside the scope of what Google can accomplish with voice search because it's going to be complex. There's going to be value in multiple results then so that's [inaudible 00:19:26.1]. Mark: Okay well you know I think that's a fair answer. I think when you ask Google a question, if it's a quick answer like time and date that makes sense. But if you're asking how to replace a sprinkler head, Google might give you a short response but you're not looking for a three step process for that. You're probably looking for pictures or video or more in depth you know of your in-depth guide. And so getting that response is actually a good thing, getting that being that featured response at Google will probably be a good thing because more people are going to click through to your page right? Corey: Yup and there's also still value being lost right now from what they call no click searches. Where maybe you appear within the knowledge card, like the top of the results; people see your brand, they see that it's from you; they don't click through and see your analytics. But at some point who cares, if they still saw your brand you still helped them, and they still see then you may have accomplished what you set out to you anyway. It's just not going to be attributed as well. Mark: Right; of course. I get it the top of funnels sort of just brand awareness and awareness to your brand, what's better to vouch for you than Google right? Corey: Right. Mark: If Google's going to feature you on their search result page that's a pretty good thing and if people don't know what I'm talking about here do you have an example that you know off the top your head where a knowledge card will show up. Corey: Recipes are a big one now. I don't know any exactly at the top of my head but- Mark: Well didn't … wasn't there that one for a while which was why are fire engines red; do you remember seeing that? Corey: I think that so, it sounds familiar. Mark: Yeah if people haven't looked at this, do a search for why are fire engines red and take a look at what the response is. They may have updated it since I last did it but it was just somebody had the game of the knowledge card and it was kind of a crazy response. Corey: Yeah, but it's still not that hard to do. Mark: So I … okay so if I already had, you can't leave with that not go into it. Corey: Well we know what their data sources are so yeah you can … what Google is not good at is understanding what you're telling it, and that's what they're working on back checking right? Being able to actually understand is this good and not are these words here and phrase didn't maybe kind of mean something. And that's what I think they'll improve that maybe next. I think it'll take a while because they're still behind what a lot of the articles give them credit for now if you like but we do know which way they're going. There's an analogy. I love Aj Kohn as an SEO blogger; his company is called The Blind Five Year Old because that's how he perceives Google still. Kind of hyperactively bouncing out of your sight not really knowing what it's doing but they are still moving in a direction that makes sense. And with the knowledge cards there's different sources of data where I mean you can literally just put it in and hope that Google crawls it. So you can update Wikidata at wikidata.org put in some bunk information and sure they might index it. They might see that it's not that hard to fool it. Mark: That's funny. All right well I got some pressing question to get to here. Even though this is fascinating and would be fun to explore all the idiosyncrasies of the world of Google but let's talk about, let's put ourselves in a position of a buyer looking to acquire a business and I want to have more opportunities for SEO and how to uncover some of those and where some of the mistakes are. But before we do that let's talk about due diligence side of things. So he's looking to buy a business, it receives a good amount of traffic from natural organic rankings. What are some of the things that people should be looking out for when doing due diligence? For example private blog networks, are these still something to look out for or are there other things that you may want to caution people on inspecting before they do an acquisition? Corey: Right so without a doubt I would never buy any website without looking at its backlink portfolio. There are basically two arms of SEO right, you've got what happens on the website and off of it. I'm not so concerned about what's happening on the website. I know just based on my background I can probably make it a lot better. But I know that it's not a danger zone, the links are. So first are they trending upwards that's a good sign; bad links tend to get moderated. It makes sense, if somebody spams a whole bunch of forums or blogs they use a piece of software, it's going to get turned away and their Google patents that talk about this as a signal. Like if somebody blasts 100,000 links and all of a sudden they disappear I immediately know something's wrong. And even if something's not wrong if they had a good reason for that to happen, I still haven't really seen one, but if they did it … that pattern looks really bad. So that's the first thing, okay I guess I start to dig into it and I start to look for schemes like you mentioned; are there link wheels, are there … you mentioned that you and I are pre-Panda people a little bit here. I've … I know the schemes because I've used the schemes. I've tried the schemes and I know what all of them look like. There's any of maybe a dozen which might go beyond our time right now but- Mark: So with some of these schemes how would somebody identify these? Obviously link patterns so seeing declining back wings would be an example of things being moderated away from low quality sites or even high quality sites where it's been spammed to a public place. But for like a Link Wheel or a PBN, are there tools that you would recommend somebody use for this or is it really just something where you need to hire somebody like you to be able to help identify these schemes? Corey: Sure. Well I won't go so far as to say someone has to hire me but I do have a lot of skepticism in the tools only because we see them throw a lot of false positives. They do good things too but I've got a team that's used every backlink auto link tool I think at this point and they're flawed certainly. Especially when you pair them with the activity of disavowing links which is usually the natural next step. When you find bad links people tend to use the disavow form in Google Search Console and that's irreversible so it's really really dangerous. We've very frequently been approached by people that ran an automated link audit, got a lot of terrible advice, disavowed a lot of good links, their rankings went away, and they need help and all we can say is well now you've just got to rebuild like you shouldn't have done this. That was a bad idea. So I think it's just about recognizing the schemes and the most overarching witness test in my mind is does this double as good marketing. Sometimes it's just a completely automated site like you see a lot of these like statistics websites, and he ways websites, those big automated plays. I would usually say if a site links to every site on the internet which you can usually see, like is it linking to every domain alphabetically; you see that a lot on the backlink tools. I don't worry about those. I don't think you should disavow those because that's not a scheme. That's not a pattern that you want out and will and that's a flaw in every auditing tool I've used. So I wouldn't worry about those. I also wouldn't worry about anything that is editorially relevant. Like is it editorial, a guest post, a press release, a mention of any kind really that happened from a human but if it didn't and you can usually tell by this kind of thumbing through the side a little bit that usually means that your link is appearing besides other schemes. And if a link is really easy to get that by definition kind of makes it a bad link which is counterintuitive right? You've got all these SEO services that are offering fast easy links for everybody. That's flawed because if it's really for anybody that means that you're link appears besides porn sites you know fill affiliates like all sorts of really kind of sketchy looking stuff. It shouldn't be easy for everybody and that's really the way to tell it I think. Mark: So something that we see with Quiet Light Brokerage in our backlink profile is we'll get a piece published informs or entrepreneur or in [inaudible 00:27:36.5 a good piece and obviously we love those backlinks. But then sure enough there is these really low quality sites that will take that article that blog post and they'll republish it and you know it's just a complete spammy site. You can tell that there's never a human that has touched that site other than initially [inaudible 00:27:55.7]. Are those backlinks, if somebody is doing a backlink analysis on the site and they see some good high quality backlinks but then they see a whole bunch of copycats stuff is that anything to worry about in your opinion? Corey: It depends a little bit on the site. If they're purely just scraping forbes, I'd say well today link back because if they do it reminds me a little bit of press indication which is perfectly natural and it's a signal that I think any grown up brand is going to have. Like you've got basically every publicly traded companies running out regular press releases so if I put on my … like if I'm Google Ad that actually looks okay. But if it's a really low quality site you might see them also doing other shady stuff so you might have to look at their backlink portfolio and kind of pick apart what they're doing. Mark: Okay fair enough that is good advice. And if anybody is listening to this and you're completely lost as far as what Corey is talking about here I'm sure you could reach out to him and get a little bit more insight into some of these things. The world of SEO is kind of this big old rabbit hole, you can understand on a very basic level or you can get into [inaudible 00:29:02.3] sort of the more nuanced stuff. In which case you're talking about link wheels and different types of shapes as far as linking patterns which I've thrown most of that out the window years ago when I started seeing a lot of the updates come through. So and I want to talk about that you know we talked a lot about backlinks and backlink profiles, it's been my perception and please correct me you are the expert in this not me, it's my perception that backlinks haven't been so much devalued as might have been surpassed or might have … might be having other ranking indicators kind of come up alongside backlinks as being important. And one of the ones I've seen has been topic coverage, topical coverage on a page. So an example of that would be we have a blog post on I want the Seller's Discretionary Earnings well we also want to cover not only what a Seller's Discretionary Earnings but what does it mean for an Accrual Basis versus Cash Basis Accounting and you know what is Net Income, what is Gross Revenue because these are related topics to the one thing so having all that content now is a good signal to Google. In my correct or incorrect or off based when I say that the backlinks while still important are playing alongside some of these other newer ranking factors? Corey: Yeah I mean you'd be correct in saying On Page matters more and in more nuanced ways. It's tough to weigh like do links matter less because links are infinite really. On page is still finite so I think in that math equation links can never matter less because you can always do more with links. You can't always do more with your site so that makes that equation interesting. But yeah since the Hilltop Algorithm which I believe was written by Krishna Bharat, he published a paper that it's actually really old but it was pre-Panda by a lot and it broke down I think what they've been building upon for a really long time which for the first time defined what they call topical experts. And if you really dig into the paper it appears to be talking about domains as experts and they played but there's a little bit you know you had authorship of Google+ I think was one sort of riff on that idea of trying to figure out who really knows about a topic. And around that time SEO's like crazy with the concept of relevance. People are saying well you only want links from relevant sites. I think that's bunk because well do I not want to link from CNN they don't … they're not an SEO website, obviously I do, obviously that's still a good link. But there's more value if I get somebody from within my space on average. So it's just one more metric, it's a little bump I feel like in their favor if they're relevant or if you're relevant. They're looking at the themes throughout your site definitely. So to your point yeah that exact same idea, the more you cover a topic the more I think your ratio of expertise is strengthened there. And for the same reason Mike & Mitch E-commerce Shop should absolutely be able to outrank Amazon. They're generic, they don't have that focus and we see that a lot. Mark: That seems to be a recurring theme of this podcast here; how to be Amazon at their own game. And I've talked to so many e-commerce business owners who get frustrated by … when they put their own listings up on Amazon and all of a sudden Amazon's outranking their niche store. But I think your point of if you have good topical coverage on your site, if you're doing … if you're making sure the on page is right you should be able to outrank Amazon because it is a specialized site. And that actually said was a really nice link into the final section I want to cover and that is opportunities for pretty much any buyer when you're looking to acquire an online business opportunities in SEO. I see huge opportunities with most of the stuff I look at and working on the on page SEO, what are a few areas in your opinion that people should look at when they're looking to grow the SEO presence of either an e-commerce shop or a SAS business or a content business but really kind of looking at that a SEO portion, what are a few areas that are common pieces of low hanging fruit that you see? Corey: Sure. Well since Panda there are a lot more diamonds in the rough I feel like that just have broken on page SEO and the poster case study going all the way back was Danny Webb right? Everyone was talking about Danny Web which was one of the biggest tech forms, they lost easily all their rankings when Panda first hit and they managed to recover by removing what people later called thin content. Which were just pages that might have fifty words on the page, it was all the different individual profiles that people had, there were millions of them. Most of them were a bad search engine experience. So when I see a site that has a lot of pages that don't offer value to Google but don't carry the no index tag, the media tag, and the source code [inaudible 00:33:56.5] but code in the source that says keep us out of Google's index. I know well hey I can do that and overnight strengthen the stuff I want to keep and cut out the stuff that's just never going to be of any value and that's going to help a lot. I also look for sites that don't have a keyword strategy, sites that for whatever reason have never had any links but still enjoyed some organic success. There's a lot of ways to play this. In total there's I think a couple hundred ranking factors. I basically just look for a couple that have been 100% neglected because I feel like that's where people leave the most money on the table basically where I can see a quick one. Mark: Yeah, I think again coming back to round out this discussion, I think since the updates and after the updates everybody was scared of organic traffic and understandably so. I mean it was very difficult for a lot of people because they owned a business overnight an update happened and the rankings are gone, revenue is gone. A lot of businesses were built on this kind of shaky SEO and Google's done a good job of cleaning that up. But people now see the benefit of relying on paid acquisitions. As a result though I think there's a huge opportunity for buyers to take a look at pretty much any property that is not on Amazon. So any web-based property content sites etcetera etcetera and be able to really grow that business through good SEO practices. As you said looking at keyword strategies are, is there any keyword strategy there, or do they have good topical coverage, are they doing the basics to be able to rank well, and because no one is really doing that or very few people are really doing that on page SEO anymore it's kind of amazing how quickly it's fallen out of favor. Yeah so let me ask you if anybody wants to talk to you what's the best way to reach you? Corey: Sure. Well they can pop on our website which is just northcutt.com drop me an email it's just corey@northcutt.comor follow me on Twitter corey_northcutt to be my first name, any of those work. Mark: All right good. We'll all link to those in the podcast page the show notes so everyone can take a look at that. And you know again we don't get kickbacks from guests or anything like that but we do refer people that we've used in the past successfully and the services of yours is definitely a service that I've gotten good value out of. I know you did some work for me, I think it was back in October your group did some work for me and those pages are doing quite well now so thank you for that. I never gave you an update on that; they're doing pretty well. Corey: Yeah, sure. Mark: Yeah so thanks so much for coming on. I think this is an interesting topic and maybe one that we need to explore again in the future. Corey: Oh I'd like that. Mark: All right, thanks Corey. Corey: Thanks Mark.

Tech Policy Podcast
#113: Wikipedia for Data

Tech Policy Podcast

Play Episode Listen Later Aug 12, 2016 25:38


Most people are familiar with Wikipedia, but there's a lot more to "open data" than the convenience of checking how tall your favorite Olympic athlete is. Open databases can play a key role in supporting research and innovation, but they also raise questions about intellectual property and fair compensation for creators. How are databases like Wikidata regulated in Europe, and how does that approach differ from the U.S.? Julia Schuetze, a Euromasters student and tech strategist at Wikimedia Germany joins the show.