-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Initial high-level goals and outline #2
Comments
Just some links to a few other array-based file formats that may be of interest: |
OK am back from vacation and ready to rumble. made some child issues off this one to start tracking different pieces and make 'threads' for discussing those sub-points |
Sry have been preoccupied with events on campus, will be returning to this next week. Numpydantic is near a 1.0 except for tests and docs, so its The intro to the paper can happen async, but once we have those pieces in place we can do the meat of the results |
hey every1. i just "officially" released numpydantic, so the next step now is to put that in linkml arrays generator (should only take ~a day) and then rework unfortunately, my employer has decided to commit egregious unfair labor practices in the form of police violence against my students and colleagues, so starting next week I will be on strike and not doing any work that brings any benefit to my employer - and unfortunately my academic work is decidedly within the scope of struck work. The strike will last as long as June 30th when the grad student contracts expire, but may end sooner than that, pretty unclear at this point. hope u all understand |
back at work and ready to roll. got numpydantic linkml PR open linkml/linkml#2178 and about to update my_dataset: !core.nwbfile.NWBFile
# ... various metadata fields
acquisition:
probe_0_lfp: !core.ecephys.LFP
probe_0_lfp_data: !core.ecephys.ElectricalSeries
data:
# ... metadata attributes
array: # specify using a relative path and hash
path: probe_0_lfp_data.zarr
hash:
value: # some long hash
type: blake2b
electrodes:
table: !reference /general/extracellular_ephys/electrodes
array: [0, 1, 2, 3, 4, 5] # inline arrays should work the same as path references
general:
extracellular_ephys:
electrodes:
# ... and so on
stimulus:
presentation:
my_stimulus_video: !core.base.TimeSeries
data:
array:
path: my_video.mp4 # videos behave the same as arrays
hash: # ... in some serialized form, just as an example. which is sort of like what's going on in |
sneek preview of the schema models https://github.com/p2p-ld/nwb-linkml/tree/linkml-arrays/nwb_linkml/src/nwb_linkml/models/pydantic/core that was super easy. need to make some changes in upstream linkml and i think i can scrap pretty much all of my monkeypatching of the generator too. then it's just a matter of writing dumpers and loaders. one of those times where you thank yourself for overengineering something before bc it was extremely simple to just swap out the translation/generation routine here. i think u all are gonna get a kick out of nwb-linkml once it's working by how simple it is |
Alright my dogs after a long delay, i'm gonna enter into writing mode on this starting wednesday. My goal (!!) is to shoot for 2 week writing sprint to get a draft in place, and i'm going to be doing additional demos and proofs of concept as i go. What I want to shoot for
but as always this is just what's going into my draft, and so if y'all want to add other things then by all means. I'm going to be doing my own limited bit of history but I think @cmungall and @oruebel your input on the history of arrays in linked data as well as the need for the hdmf-schema and hdmf would be super valuable here :) lets get this show on the road !!!!! |
Great! November for me has unexpectedly filled up with grant writing, but I will try to help with writing and reviewing where I can. |
no problem. reading back the initial issue, i realize as i am drafting that i am mixing in some of the nwb stuff just because it's part of the same thought to me. i figure since we were planning on twinned pieces that we can rearrange and split the words on the page once they are there if we want to, but i am starting with the stuff that would be for the perspective piece re: the schema-centric linkml-arrays approach to a data standard |
Goal: A slightly technical perspective piece describing the problem of representing arrays in schema and how array support within LinkML solves the problem
A rough first outline:
Potential target journals:
We would write a second paper on LinkML arrays for NWB/neurophysiology, more for a neuroscience audience.
The text was updated successfully, but these errors were encountered: