No need for DatabaseCleaner (rolling back transactions are usually faster than truncate).
- Jun 2020
-
github.com github.com
-
-
exts.ggplot2.tidyverse.org exts.ggplot2.tidyverse.org
-
Ggplot2 extensions. (n.d.). Retrieved June 14, 2020, from https://exts.ggplot2.tidyverse.org/
-
-
yalantis.com yalantis.com
-
Consider the following when picking technologies
-
-
-
Guardian, C. J., Melissa Bailey, Danielle Renwick, The. (2020, June 6). Exclusive: Nearly 600 — And Counting — US Health Workers Have Died Of COVID-19. Kaiser Health News. https://khn.org/news/exclusive-investigation-nearly-600-and-counting-us-health-workers-have-died-of-covid-19/
-
-
surgisphere.com surgisphere.com
-
Desai, S. (2020, May 29). Response to Widespread Reaction to Recent Lancet Article on Hydroxychloroquine. Surgisphere Corporation. https://surgisphere.com/2020/05/29/response-to-widespread-reaction-to-recent-lancet-article-on-hydroxychloroquine/
-
-
www.thelancet.com www.thelancet.com
-
McDonald, E. S. (2020). COVID-19 and essential pregnant worker policies. The Lancet Infectious Diseases, 0(0). https://doi.org/10.1016/S1473-3099(20)30446-1
-
-
dev.mysql.com dev.mysql.com
-
Deadlocks are a classic problem in transactional databases, but they are not dangerous unless they are so frequent that you cannot run certain transactions at all. Normally, you must write your applications so that they are always prepared to re-issue a transaction if it gets rolled back because of a deadlock.
-
-
api.rubyonrails.org api.rubyonrails.org
-
transaction calls can be nested. By default, this makes all database statements in the nested transaction block become part of the parent transaction. For example, the following behavior may be surprising: User.transaction do User.create(username: 'Kotori') User.transaction do User.create(username: 'Nemu') raise ActiveRecord::Rollback end end creates both “Kotori” and “Nemu”. Reason is the ActiveRecord::Rollback exception in the nested block does not issue a ROLLBACK. Since these exceptions are captured in transaction blocks, the parent block does not see it and the real transaction is committed.
How is this okay??
When would it ever be the desired/intended behavior for a
raise ActiveRecord::Rollback
to have absolutely no effect? What good is the transaction then??What happened to the principle of least surprise?
Is there any reason we shouldn't just always use
requires_new: true
?If, like they say, the inner transaction "become[s] part of the parent transaction", then if anything, it should roll back the parent transaction too — not roll back nothing.
-
One workaround is to begin a transaction on each class whose models you alter:
-
- May 2020
-
github.com github.com
-
Deepset-ai/haystack. (2020). [Python]. deepset. https://github.com/deepset-ai/haystack (Original work published 2019)
-
-
covid.deepset.ai covid.deepset.ai
-
Corona Scholar: Scientific COVID-19 Knowledge
-
-
www.ukcdr.org.uk www.ukcdr.org.uk
-
UKCDR - COVID-19 Research Project Tracker
-
-
featuredcontent.psychonomic.org featuredcontent.psychonomic.org
-
Herzog, S. (2020, May 21). Boosting COVID-19 related behavioral science by feeding and consulting an eclectic knowledge base. Psychonomic Society Featured Content. https://featuredcontent.psychonomic.org/boosting-covid-19-related-behavioral-science-by-feeding-and-consulting-an-eclectic-knowledge-base/
-
-
ncs-tf.ch ncs-tf.ch
-
Swiss national COVID-19 science task force. Policy Briefs. https://ncs-tf.ch/en/policy-briefs
-
-
-
Database of public health guidance on COVID-19. (2020 May 14). HIQA. https://www.hiqa.ie/reports-and-publications/health-technology-assessment/covid-19-public-health-guidance-database
-
-
www.essentialsql.com www.essentialsql.com
-
I think you should normalize if you feel that introducing update or insert anomalies can severely impact the accuracy or performance of your database application. If not, then determine whether you can rely on the user to recognize and update the fields together. There are times when you’ll intentionally denormalize data. If you need to present summarized or complied data to a user, and that data is very time consuming or resource intensive to create, it may make sense to maintain this data separately.
When to normalize and when to denormalize. The key is to think about UX, in this case the factors are db integrity (don't create errors that annoy users) and speed (don't make users wait for what they want)
-
Can database normalization be taken too far? You bet! There are times when it isn’t worth the time and effort to fully normalize a database. In our example you could argue to keep the database in second normal form, that the CustomerCity to CustomerPostalCode dependency isn’t a deal breaker.
Normalization has diminishing returns
-
Now each column in the customer table is dependent on the primary key. Also, the columns don’t rely on one another for values. Their only dependency is on the primary key.
Columns dependency on the primary key and no dependency on other columns is how you get 2NF and 3NF
-
A table is in third normal form if: A table is in 2nd normal form. It contains only columns that are non-transitively dependent on the primary key
3NF Definition
-
- Apr 2020
-
stackoverflow.com stackoverflow.com
-
web.archive.org web.archive.org
-
From a narratological perspective, it would probably be fair to say that most databases are tragic. In their design, the configuration of their user interfaces, the selection of their contents, and the indexes that manage their workings, most databases are limited when set against the full scope of the field of information they seek to map and the knowledge of the people who created them. In creating a database, we fight against the constraints of the universe – the categories we use to sort out the world; the limitations of time and money and technology – and succumb to them.
databases are tragic!
-
-
psyarxiv.com psyarxiv.com
-
De Brier, N., Stroobants, S., Vandekerckhove, P., & De Buck, E. (2020, April 23). Factors affecting mental health of health care workers during coronavirus disease outbreaks: a rapid systematic review. https://doi.org/10.31234/osf.io/w9uxs
-
-
en.wikipedia.org en.wikipedia.org
-
columnar databases are well-suited for OLAP-like workloads (e.g., data warehouses) which typically involve highly complex queries over all data (possibly petabytes). However, some work must be done to write data into a columnar database. Transactions (INSERTs) must be separated into columns and compressed as they are stored, making it less suited for OLTP workloads. Row-oriented databases are well-suited for OLTP-like workloads which are more heavily loaded with interactive transactions. For example, retrieving all data from a single row is more efficient when that data is located in a single location (minimizing disk seeks), as in row-oriented architectures. However, column-oriented systems have been developed as hybrids capable of both OLTP and OLAP operations, with some of the OLTP constraints column-oriented systems face mediated using (amongst other qualities) in-memory data storage.[6] Column-oriented systems suitable for both OLAP and OLTP roles effectively reduce the total data footprint by removing the need for separate systems
typical applications (adding new users data, or even retrieving user data) are better done in (standard) row-oriented DB. Typical analytics application, such as even simple AVG(whole column) are much slower because the elements of the same column are stored far away from each other in a traditional row-oriented DB, hence increasing disk-access time.
-
seek time is incredibly long compared to the other bottlenecks in computers
-
Operations that retrieve all the data for a given object (the entire row) are slower. A row-based system can retrieve the row in a single disk read, whereas numerous disk operations to collect data from multiple columns are required from a columnar database.
Tags
Annotators
URL
-
-
covid19.scops.ai covid19.scops.ai
-
Scops. Topic Stream Graph. Covid19.scops.ai. https://covid19.scops.ai/superset/dashboard/streamgraph/.
-
-
www.mpg.de www.mpg.de
-
Interview: Calovi, D. by Herrmann Tobias. (2020 April 03). "We don't just want to sit on the couch". Max-Planck-Gesellschaft. https://www.mpg.de/14648916/daniel-calovi-crowdfight-covid19.
-
-
www.thelancet.com www.thelancet.com
-
Viner, R. M., et al. (2020 April 06). School closure and management practices during coronavirus outbreaks including COVID-19: a rapid systematic review. The Lancet. DOI: https://doi.org/10.1016/S2352-4642(20)30095-X.
-
-
www.parliament.uk www.parliament.uk
-
Parliament. COVID-19 outbreak expert database. Parliament.uk. https://www.parliament.uk/covid19-expert-database
-
-
twitter.com twitter.com
-
ReconfigBehSci en Twitter: “‘Proper science without the drag’ – Move to the medical model of journal review: ‘Yes/No’ decision. We suggest the temporary adoption of this model for crisis-relevant material by journals. [happening already, but potentially even better models: @Meta_psy and @F1000Research?]” / Twitter. (n.d.). Twitter. Retrieved April 15, 2020, from https://twitter.com/scibeh/status/1242094075312046082
-
-
sciencebusiness.net sciencebusiness.net
-
Science|Business Database: Coronavirus Funding Opportunities. (n.d.). Science|Business. Retrieved April 20, 2020, from https://sciencebusiness.net/sciencebusiness-database-coronavirus-funding-opportunities
-
-
dba.stackexchange.com dba.stackexchange.com
-
Relational databases are designed around joins, and optimized to do them well. Unless you have a good reason not to use a normalized design, use a normalised design. jsonb and things like hstore are good for when you can't use a normalized data model, such as when the data model changes rapidly and is user defined. If you can model it relationally, model it relationally. If you can't, consider json etc.
-
Joins are not expensive. Who said it to you? As basically the whole concept of relational databases revolve around joins (from a practical point of view), these product are very good at joining. The normal way of thinking is starting with properly normalized structures and going into fancy denormalizations and similar stuff when the performance really needs it on the reading side. JSON(B) and hstore (and EAV) are good for data with unknown structure.
-
- Mar 2020
-
clickhouse.tech clickhouse.tech
-
github.com github.com
-
docs.google.com docs.google.com
-
I chose all my scholarly journals, I put them together. I chose some YouTube videos; they were –IF: Mm-hmm.CF: – like, a bunch of TED talks.
Compiling research materials.
Is there room for us to think about the iterative process; can we work with instructors to "reward" (or assign) students to alternate the searching, reading and writing.
-
And – And I seen how – I saw how many, um, scholarly journals or how many sources came up for it, right? Um, number of sources. Right. And then, if I – if I felt like it wasn’t enough for me to thoroughly talk about the topic, I would move on. Right? So, when I did segregation, there – like, I guess, like, my specific topic was modern-day, so there wasn’t really much about it. Right? So, not much info. Right? And then, when I did gentrification, there were a lot, right?
This part of the process is interesting to me. Links topic selection to search (seemingly a single search).
It also seems a little misguided. What can we do in our lessons that could make tiny changes to this attitude?
-
-
-
unidad_COVID2019,pais
-
- Oct 2019
- Sep 2019
-
web.archive.org web.archive.org
-
The problem with the annotation notion is that it's the first time that we consider a piece of data which is not merely a projection of data already present in the message store: it is out-of-band data that needs to be stored somewhere.
could be same, schemaless datastore?
-
many of the searches we want to do could be accomplished with a database that was nothing but a glorified set of hash tables
Hello sql and cloure.set ns! ;P
-
There are objects, sets of objects, and presentation tools. There is a presentation tool for each kind of object; and one for each kind of object set.
very clojure-y mood, makes me think of clojure REBL (browser) which in turn is inspired by the smalltalk browser and was taken out of datomic (which is inspired by RDF, mentioned above!)
-
- May 2019
-
about.jstor.org about.jstor.org
Tags
Annotators
URL
-
- Oct 2018
-
www.nature.com www.nature.com
-
Massive mining of publicly available RNA-seq data from human and mouse
-
- Sep 2018
-
www.ncbi.nlm.nih.gov www.ncbi.nlm.nih.gov
-
ATtRACT-a database of RNA-binding proteins and associated motifs
Tags
Annotators
URL
-
- Apr 2018
-
www.sarahmei.com www.sarahmei.com
-
The takeaway from the article: Choose document-oriented database only when the data can be treated as a self-contained document
-
- Dec 2017
-
openspeakers.org openspeakers.org
-
Find Open Speakers
Tags
Annotators
URL
-
-
alleledb.gersteinlab.org alleledb.gersteinlab.orgAlleleDB1
-
AlleleDB is a repository, providing genomic annotation of cis-regulatory single nucleotide variants (SNVs) associated with allele-specific binding (ASB) and expression (ASE).
Tags
Annotators
URL
-
- Nov 2017
-
stackoverflow.com stackoverflow.com
-
select top 1 * from newsletters where IsActive = 1 order by PublishDate desc
This doesn't require a full table scan or a join operation. That's just COOL
-
-
www.datavisor.com www.datavisor.com
-
They have a very simplistic view of the activity being monitored by only distilling it down into only a few dimensions for the rule to interrogate
Number of dimensions need to be large. In normal database systems these dimensions are small.
-
- Aug 2017
-
niemanreports.org niemanreports.org
-
Football Leaks, which consists of 1.9 terabytes of information and some 18.6 million documents, ranging from player contracts to emails revealing secret transfer fees and wages, is the largest leak in the history of sport.
"Football Leaks, which consists of 1.9 terabytes of information and some 18.6 million documents, ranging from player contracts to emails revealing secret transfer fees and wages, is the largest leak in the history of sport."
A pity this information is not available to the public.
Given the limited release of documents, is it really the largest leak in the history of sport?
The ICIJ offshore database may not be complete, but there is at least something, and it is searchable.
Hopefully EIC will also follow this example.
-
- Jun 2017
-
wiki.opencog.org wiki.opencog.org
-
The vertices and edges of a graph, known as Atoms, are used to represent not only "data", but also "procedures"; thus, many graphs are executable programs as well as data structures.
Rohan indicated that procedures are also part of the graph. let us find out why.
Tags
Annotators
URL
-
- May 2017
-
static1.squarespace.com static1.squarespace.com
-
Databases tend toward inclusivity, narratives toward selectivity
This is a helpful little encapsulation of the central problem in the debate.
-
- Mar 2017
-
med.stanford.edu med.stanford.edu
-
Genome Sequence Archive (GSA)
Database URL is here: http://gsa.big.ac.cn/
Note: metadata is INSDC format, but this database isn't part of the INSDC, so you'll still need to submit your data to one of those databases to meet internationally recognised mandates
-
- Feb 2017
-
wiki.dbpedia.org wiki.dbpedia.orgDBpedia1
Tags
Annotators
URL
-
- Jan 2016
-
ifsacop21.wordpress.com ifsacop21.wordpress.com
-
international law system
Wikipedia page https://en.wikipedia.org/wiki/International_legal_system
Does an International legal informatics database exist yet?
-
- Mar 2015
-
www.image-net.org www.image-net.orgImageNet1
-
ImageNet is an image database organized according to the WordNet hierarchy (currently only the nouns), in which each node of the hierarchy is depicted by hundreds and thousands of images. Currently we have an average of over five hundred images per node. We hope ImageNet will become a useful resource for researchers, educators, students and all of you who share our passion for pictures.
Tags
Annotators
URL
-
- Dec 2014
-
opengtindb.org opengtindb.org
-
- Sep 2014
-
inst-fs-iad-prod.inscloudgate.net inst-fs-iad-prod.inscloudgate.net
-
Most systems av ailable today use a single database.
Even after all these years...this is essentially still true.
-
- Jan 2014
-
onlinelibrary.wiley.com onlinelibrary.wiley.com
-
The initial inputs for deriving quantitative information of gene expression and embryonic morphology are raw image data, either of fluorescent proteins expressed in live embryos or of stained fluorescent markers in fixed material. These raw images are then analyzed by computational algorithms that extract features, such as cell location, cell shape, and gene product concentration. Ideally, the extracted features are then recorded in a searchable database, an atlas, that researchers from many groups can access. Building a database with quantitative graphical and visualization tools has the advantage of allowing developmental biologists who lack specialized skills in imaging and image analysis to use their knowledge to interrogate and explore the information it contains.
1) Initial input is raw image data 2) feature extraction on raw image data 3) extracted features stored in shared, searchable database 4) database available to researchers from many groups 5) quantitative graphical and visualization tools allow access to those without specialized skill in imaging and image analysis
-