Closed FredericBlum closed 1 year ago
Our current workflow is (we used this on the Andean data, if I remember properly):
We can discuss to make a command in lexibank that leads to the automatization of step 2, I think we may evan have an issue on this.
We could, of course, also discuss initiating individual profiles for individual languages by modifying the init-profile command.
In my experience, however, this two-step workflow is easier, especially when there are more languages in a dataset.
@Tarotis, once you are ready with some data, please get back to me, and we write this up for inclusion in pylexibank.
I may use this to teach you how to write lexibank commands.
@LinguList As I have a first set of items ready for all languages and am only filling gaps right now, we can start working on the ortography. Do you have any example cases (+ commands) where you used this workflow that I could adapt to blumpanotacana?
@LinguList Creating language-specific profiles by adding a lexibank-command would be a great next task to get some code review and feedback on package development. Maybe we can set this up for the second week of March?
Let me check for the code now.
Let me first tell you the strategy (which is important):
etc/orthographies/language_id.tsv
In a package, you would access the cldf code, not any single profile, create the language-specific profiles from that and write them to the files in etc
or maybe a user-specified folder.
I applied this code the first time for the lsi-project. I also shared it with Sandra, who used it in her Mixtecan study (at least at some point, she never gave me feedback on that). I'd start from the code in LSI, but note that some aspects of the code are specific to the dataset, so you don't want to use them.
If you want as an alternative an init_profile function that does the init-profile and write multi-language profiles, this can of course also be done, but it may result in extra work, I think. Yet, of course, you can now check profiles for individual languages, so it would not be too bad. Let me know what you prefer.
Thanks, the script you provided worked perfectly, with only very minor modifications necessary
Hello everybody,
I am wondering right now how I would specify
lexibank.init_profile
to distinguish between the different languages in my dataset. I have seen two structures so far in Lexibank sets: a) Individual ortho-profiles for each language, or b) a column specifying language handles for which a certain mapping applies. Can I create either of these, but preferably a), usinginit_profile
?I saw there are some hints in the code, but it is not clear to me what
args.context
is about and how I would specify the command line command to use this.Looking forward to hearing your opinions on this.