LinguaLibre

Difference between revisions of "Chat room/Archives/2022"

< LinguaLibre:Chat room

(Archiving posts older than 6 months)
 
Line 8: Line 8:
  
 
== Merging of items about languages ==
 
== Merging of items about languages ==
 +
:''{{Done}} final fix for Chinese writing, Duala, Mossi on 07:01, 24 December 2023 (UTC) by [[User:Dragons_Bot]]/[[USer:Yug]].''
 
:''See also [[Help:SPARQL]] and [[Help:SPARQL for maintenance]].''
 
:''See also [[Help:SPARQL]] and [[Help:SPARQL for maintenance]].''
 
Hi y'all,
 
Hi y'all,
Line 36: Line 37:
  
 
Cheers, [[User:VIGNERON|VIGNERON]] ([[User talk:VIGNERON|talk]]) 09:29, 19 February 2022 (UTC)
 
Cheers, [[User:VIGNERON|VIGNERON]] ([[User talk:VIGNERON|talk]]) 09:29, 19 February 2022 (UTC)
 +
[[File:Dragons bot on babel tower.jpg|Thumb|300px|{{u|Dragons Bot}} monitoring Lingualibre.]]
 
:Thanks VIGNERON for finding them and cleaning it. Now what to do with recording items that use the doublon language item (for example with [[Special:WhatLinksHere/Q52071|Duala]]). I think we must modify {{P|4}} for all recording items so that languages are not counted twice and also to clean up the database (there are also transcription problems for items listed in the Duala example). [[User:Pamputt|Pamputt]] ([[User talk:Pamputt|talk]]) 16:16, 19 February 2022 (UTC)
 
:Thanks VIGNERON for finding them and cleaning it. Now what to do with recording items that use the doublon language item (for example with [[Special:WhatLinksHere/Q52071|Duala]]). I think we must modify {{P|4}} for all recording items so that languages are not counted twice and also to clean up the database (there are also transcription problems for items listed in the Duala example). [[User:Pamputt|Pamputt]] ([[User talk:Pamputt|talk]]) 16:16, 19 February 2022 (UTC)
 
::Thank you {{ping|VIGNERON}} for pointing these out. As you can see, most of them were not created manually but using the tool (the pages wheighted circa 4kB, with labels in many languages). It seems that the Lingua Importer tool has (or had?) a problem, but I could not reproduce it (trying to import languages that are already in LL wikibase).<br/> During last summer's hackathon we talked a bit about languages in our wikibase, but I can't remember why we need to have language elements in our Wikibase, and not just use the existing base of WikiData 🤔 — '''[[User:WikiLucas00|WikiLucas]]''' [[User talk:WikiLucas00|(🖋️)]] 23:23, 19 February 2022 (UTC)
 
::Thank you {{ping|VIGNERON}} for pointing these out. As you can see, most of them were not created manually but using the tool (the pages wheighted circa 4kB, with labels in many languages). It seems that the Lingua Importer tool has (or had?) a problem, but I could not reproduce it (trying to import languages that are already in LL wikibase).<br/> During last summer's hackathon we talked a bit about languages in our wikibase, but I can't remember why we need to have language elements in our Wikibase, and not just use the existing base of WikiData 🤔 — '''[[User:WikiLucas00|WikiLucas]]''' [[User talk:WikiLucas00|(🖋️)]] 23:23, 19 February 2022 (UTC)
 +
:Hello {{Ping|WikiLucas00|VIGNERON|Pamputt}}
 +
:{{Done}} Issue solved. After 1+ year monitoring of the issue, learning bots ({{u|Dragons_Bot)), and 6+ hours coding, all existing recordings with
 +
:* erroneous [[Q130]] (Chinese writing),
 +
:* erroneous [[Q52071]] (Duala)
 +
:* erroneous [[Q170137]] (Mossi)
 +
:were edited to point with the correct value. On both Lingualibre and Wikimedia Commons. So there is no more remaining duplicated languages.
 +
:Wishing everyone a good Xmas season ! [[User:Yug|Yug]] ([[User talk:Yug|talk]]) 07:01, 24 December 2023 (UTC)
  
 
== MediaWiki customizations of LinguaLibre ==
 
== MediaWiki customizations of LinguaLibre ==

Revision as of 07:01, 24 December 2023

Comment changer de pseudonyme

Bonjour, sur les projets de Wikimedia, mon pseudonyme est Manjiro91 (et anciennement GamissimoYT), comment change-t-on de pseudonyme ? GamissimoYT (talk) 17:13, 11 January 2022 (UTC)

Bonjour GamissimoYT. Lingua Libre utilise le même pseudo que celui qui est en utilisation sur Wikimedia Commons. Donc si vous voulez utiliser le pesudonyme Manjiro91, déconnectez-vous de Lingua Libre, puis de Wikimedia Commons. Ensuite, connectez vous à Commons avec le pseudo Manjiro91 et enfin reconnectez vous à Lingua Libre. Pamputt (talk) 21:05, 11 January 2022 (UTC)

@Pamputt Mon pseudonyme Wikimedia Commons est Manjiro91 (anciennement GamissimoYT mais le changement de pseudonyme ne s'effectue pas sur LiLi. GamissimoYT (talk) 13:38, 12 January 2022 (UTC)

@GamissimoYT , tu as bien fait les connexions/déconnexions dans l'ordre que j'ai indiqué ? Si tu es sûr que tu es connecté avec Manjiro91 sur Wikimedia Commons, alors tu peux essayer de te déconnecter de Lingua Libre et te reconnecter dans la foulée. Essayer de vider le cache du navigateur peut peut-être aidé aussi. Pamputt (talk) 07:37, 13 January 2022 (UTC)

Merging of items about languages

Check-green.svg Done final fix for Chinese writing, Duala, Mossi on 07:01, 24 December 2023 (UTC) by User:Dragons_Bot/USer:Yug.
See also Help:SPARQL and Help:SPARQL for maintenance.

Hi y'all,

For the record, I just merge a couple of items about the same language:

I detected them with this SPARQL query:

SELECT ?idWD (COUNT(?item) AS ?compte) (GROUP_CONCAT(?item) AS ?items) WHERE {
  ?item prop:P2 entity:Q4 ; prop:P12 ?idWD .
}
GROUP BY ?idWD
HAVING ( ?compte > 1 )

Ping @WikiLucas00 it seems you are responsible for some of them...

Cheers, VIGNERON (talk) 09:29, 19 February 2022 (UTC) Dragons Bot monitoring Lingualibre.

Thanks VIGNERON for finding them and cleaning it. Now what to do with recording items that use the doublon language item (for example with Duala). I think we must modify language (P4) for all recording items so that languages are not counted twice and also to clean up the database (there are also transcription problems for items listed in the Duala example). Pamputt (talk) 16:16, 19 February 2022 (UTC)
Thank you @VIGNERON for pointing these out. As you can see, most of them were not created manually but using the tool (the pages wheighted circa 4kB, with labels in many languages). It seems that the Lingua Importer tool has (or had?) a problem, but I could not reproduce it (trying to import languages that are already in LL wikibase).
During last summer's hackathon we talked a bit about languages in our wikibase, but I can't remember why we need to have language elements in our Wikibase, and not just use the existing base of WikiData 🤔 — WikiLucas (🖋️) 23:23, 19 February 2022 (UTC)
Hello @WikiLucas00, VIGNERON, & Pamputt
Check-green.svg Done Issue solved. After 1+ year monitoring of the issue, learning bots ([[User:{{{1}}}|{{{1}}}]]} Languageseeker (talk) 23:17, 20 May 2022 (UTC)

Contribution: Python program to download all files created by a specific user

See also Help:Download datasets.

I wrote a python program that downloads all the files created by one user. For video files, it downloads the full webm. For audio files, the default is to download the wave file. However, for audio files, you can optionally choose either mp3 or ogg files. Currently, the configuration requires a minor modification of lluad.py. If there is strong demand, I will write a command line parser for it. Please report any bugs or errors on the github page. Feature requests are welcome. Languageseeker (talk) 02:28, 20 May 2022 (UTC)

@Languageseeker please add your tool to Help:Download datasets. It lists several tools with different specifics, your tool is welcome and may help some Python users as well. Yug (talk) 22:41, 22 May 2022 (UTC)

Garbage Values in prop:P14

See also Help:SPARQL for maintenance and Help:SPARQL_for_maintenance#.E2.9C.85_Speakers_.E2.86.92_Undefined_place_of_residence.

As part of my Anki project, I queried the entire LL database and I'm trying to parse the output of ?speaker prop:P14 ?residence. I've noticed that there are a number of garbage values in provided for P14, such as Q1, Q2, Q103962887, Q6099648, Strasbourg. There seem to be three cases.

  1. Users wishing to enter an extremely vague place such as Earth or the Universe. These should be set to None.
  2. Users accidentally linking to a disambiguation page. These require correction.
  3. Users not even entering a Wikidata item which require manual correction.

To solve the root of the problem, I propose that P14 should be restricted to only Wikidata items that exist and have P17. Languageseeker (talk) 21:22, 25 May 2022 (UTC)

@Languageseeker it's a good find. If you still have that SPARQL query under hand please add it into Help:SPARQL for maintenance. Yes, it's something we should clean up i think. There may be some few case where the speaker dont want to share its location but in 95% of cases i think we can go ahead, correct or ask them to correct it. Yug (talk) 12:39, 26 May 2022 (UTC)
I noticed that when creating a new speaker, place of learning is optional. Not cool. Yug (talk) 21:32, 27 May 2022 (UTC)
@YUG For the life of me, I can't get the federated query to work, but I have a separate query to get the location and country labels from wikidata. These are the problematic ones. Note, that Q20 is on the list because Q20 "Norway" is missing P17
  • ['MichaelSchoenitzer', None]
  • ['D.Muralidharan', None]
  • ['Kaderousse', None]
  • ['Krokus', None]
  • ['विदुला टोकेकर', 'Q103962887']
  • ['DoctorandusManhattan', 'Q2']
  • ['Justforoc', 'Q2']
  • ['Student16 de', None]
  • ['Didierwiki', 'Q6099648']
  • ['Sarah2149', None]
  • ['DomesticFrog', 'Q1']
  • ['Drkanchi', None]
  • ['Satdeep Gill', None]
  • ['Iwan.Aucamp', 'Q20']
  • ['Skimel', 'Q2']
  • ['Abeɣzan', None]
  • ['Gibraltar Rocks', None]
  • ['Bomdapatrick', None]
  • ['Ibtissam RAHMOUNI', None]
  • ['Trabelsiismail', None]
  • ['Ziko', 'Q2']
  • ['Youcefelallali', None]
  • ['Foxxipeter7', None]
  • ['Celevra089', None]
  • ['Bodhisattwa', None]
  • ['Atudu', None]
  • ['KageyamaxNishinoya', 'Q30915818']
  • ['Darkdadaah', None]
  • ['JayashreeVI', None]
  • ['रश्मीमहेश', 'Q103962887']
  • ['गीता गोविंद नेने', 'Q103893785']
  • ['Awangba Mangang', None]
  • ['Abigaljo', None]
  • ['FaelDaug', 'Q29423162']

Languageseeker (talk) 02:16, 30 May 2022 (UTC)

Anki Extension Release

I just released Lingua Libre and Forvo Addon. It has a number of advanced options to improve search results and can run either as a batch operation or on an individual note.

By default, it first checks Lingua Libre and, if there are no results on Lingua Libre, it then checks Forvo. To run as a pure Lingua Libre extension, you will need to set "disable_Forvo" to True in your configuration section.

Please reports bugs, issues, ideas on github. I would love any feedback. Languageseeker (talk) 02:23, 31 May 2022 (UTC)