Science topic

Database Management - Science topic

Explore the latest questions and answers in Database Management, and find Database Management experts.
Questions related to Database Management
  • asked a question related to Database Management
Question
18 answers
Hello,
I am trying without success to install Discovery Studio Visualizer in some laptops with Ubuntu 18.04 LTS.
Is there a workaround for install Discovery Studio Visualizer (2016 --> up) in Ubuntu 18.04 LTS?
I've tested here and it works great with Ubuntu 14, Ubuntu 16 and Debian 9...
If it's not possible to install it on Ubuntu 18, is there another software for database management? (Create and delete attributes in SDF files; sort for an specific column, etc.)
Best,
Ricardo
Relevant answer
Answer
Aiman Syafiq : I solved it in the way given below.
This might help :
first Download library (in this case libpng15)
locate the downloaded file and extract to Downloads folder.
now go to TERMINAL and type
cd Downloads/libpng-1.5.15
./configure --prefix=usr/local/libpng make check sudo make install make check
Congratulations you just installed the libpng
Now we need to create a shortcut and put it in /usr/lib for any program that is depended on it to work
In TERMINAL type:
cd (only if you are not using a new terminal) sudo updatedb (this could take a few seconds) locate libpng (locate the ".../libpng15.so.15" line and COPY)
now we will create a NEW LINK (shortcut) of that file to "/usr/lib/"
sudo ln -s /usr/local/libpng/lib/libpng15.so.15 /usr/lib/libpng15.so.15
you can now successfully execute applications that require libpng15.so.15
  • asked a question related to Database Management
Question
2 answers
Hi everyone!
What is the best way to automatically identify the material that produced an EDS spectrum? What are you favorite software tools? To be clear, I don't mean peak identification. Specifically, I want to distinguish various metal alloys.
For example, when presented with a sample of unknown material, I want to compare its spectrum to a database of standard spectra I have collected. The comparison should output some metric describing how well the specimen matches a standard in my database.
EDAX's TEAM can do this, but its database management is too cumbersome for the scale of work I want to do. Thermo Fisher's Pathfinder software can also do spectrum matching, but I don't have a license for that either. Others probably have one too. As far as I can tell, DTSA-II and Hyperspy do not have spectrum matching algorithms.
I'm aware that this technique is nuanced and has severe limitations, mainly the limits of detection for lighter elements are poor, and the quality and polish of both specimen and standard should be pristine. Even considering those criteria, this is a useful technique for my research.
Thanks in advance!
Relevant answer
Answer
Thanks, Peter. That is a really interesting paper from the lead developers on DTSA-II at NIST.
Your quote is from a section warning the reader when EDS quantification is inappropriate, specifically with severe topology.
I think the abstract summarizes the thesis of the paper really well, which is that very good accuracy can be achieved with EDS despite its reputation of being semi-quantitative.
" Electron-excited X-ray microanalysis performed in the scanning electron microscope with energydispersive X-ray spectrometry (EDS) is a core technique for characterization of the microstructure of materials. The recent advances in EDS performance with the silicon drift detector (SDD) enable accuracy and precision equivalent to that of the high spectral resolution wavelength-dispersive spectrometer employed on the electron probe microanalyzer platform. SDD-EDS throughput, resolution, and stability provide practical operating conditions for measurement of high-count spectra that form the basis for peak fitting procedures that recover the characteristic peak intensities even for elemental combination where severe peak overlaps occur, such PbS, MoS2, BaTiO3, SrWO4, and WSi2. Accurate analyses are also demonstrated for interferences involving large concentration ratios: a major constituent on a minor constituent (Ba at 0.4299 mass fraction on Ti at 0.0180) and a major constituent on a trace constituent (Ba at 0.2194 on Ce at 0.00407; Si at 0.1145 on Ta at 0.0041). Accurate analyses of low atomic number elements, C, N, O, and F, are demonstrated. Measurement of trace constituents with limits of detection below 0.001 mass fraction (1000 ppm) is possible within a practical measurement time of 500 s. "
  • asked a question related to Database Management
Question
3 answers
An Oracle server stores data logically in Database->Tablespace->Segment->Extent->Block.
I want to compare Oracle servers with MS SQL servers in terms of the logical storage of data in the database.
Relevant answer
Answer
Generally It stores data as files that each file has several records in binary format. How this files store in disk , depends on file organization that DBMS uses . For example sequential, hash and so on .
  • asked a question related to Database Management
Question
12 answers
I'm looking for a local database for managing faunistic data.
Relevant answer
Answer
Hello Dragomir;
I use Microsoft Access for my ant data (the usual assortment of information). My coauthor uses Filemaker. Access has a capacity limit while Filemaker doesn't. Importing data into Access is a pain while in Filemaker it isn't. Had I known anything when I started, I'd choose Filemaker.
Best regards, Jim des Lauriers
  • asked a question related to Database Management
Question
9 answers
I am working on my project and I have collected a bunch of data from more than 50 Summer School programs for international students in Germany. Data includes Course Content, Course duration, Course fees, Target Group, Requirement, Extra Services. Which analasis would be helpful to find out the best Summer School program?
Relevant answer
Answer
Greg Olsen Rana A. Maya Thanks a lot. Here is the final analysis of Summer Schools at Universities (University of applied sciences excluded) in Germany.
  • asked a question related to Database Management
Question
1 answer
It's the problem I have to solve, I don't have much knowledge about Distributed database, can any one help me?
Relevant answer
Answer
your Question was not clear enough so you could get help
However, I recommend you this publication of mine you might find usefull
  • asked a question related to Database Management
Question
5 answers
Dear researchers,
I've been working with a MATLAB code to run my simulation. The MATLAB code makes link with ABAQUS to run the FEM simulation and MYSQL to manage the databases for input and output files.
The present code I am working uses MYSQL to generate some tables, let's call it as input files and some other tables to save data throughout the simulation at each step.
I am wondering is there anyway to omit MYSQL, I mean is there any way to generate interactive tables in MATLAB? By interactive I mean that the table can be updated at each simulation step and the new data put in a new row in the table? So far I worked with the Table functions and I can generate the tables but the problem is that I do not know how to make it interactive or updatable!
Any idea?
Thanks a lot!
Farhad
Relevant answer
Answer
you could update a table in MATLAB using vectorization.
  • asked a question related to Database Management
Question
2 answers
Hi,
I am working on a public health research project which is based on paper based interviews. I am trying to figure out whether i can design a database in Qualtrics on to which i could enter the paper based data (myself) and export it later to a statistical program like SPSS or STATA for analysis. Far as i understand, Qualtrics is a survey tool that sends out your designed survey questionnaire to several email recipients, who fill it out and return it back to you. but can it be manipulated to work in the way i am trying to?
I am more familiar working with MS Access and Epidata when it comes to designing databases, this software is completely new to me. Any help would be greatly appreciated!
Faraz.
Relevant answer
Answer
Hi Faraz,
Michael's answer is correct. I would add that with Qualtrics you can set up a survey so that when one is completed (i.e., transfer a paper instance to a Qualtrics instance), the website will redirect to the beginning of a new survey (as if you were using a tablet device to collect responses in the field).
Hope this helps.
Kim
  • asked a question related to Database Management
Question
4 answers
The question related to automation of bibliographic information (such as author, title, subject, ISBN ISSN publisher etc. )
Relevant answer
Answer
Hi,
It seems that MARK tags.
Here are components of an academic bibliographic record:
ISBN:
0393013804
Notes:
Bibliography: p. 324-330.
Includes index.
Subjects:
Evolution (Biology)--History. Natural selection--History.
LC Classification:
QH361 .G66 1980
Dewey Class No.:
575.01/62.
  • asked a question related to Database Management
Question
26 answers
 What are best open source software available for the bioacoustic database management and for the cleaning and cutting of the raw recordings of animal sounds from the field? I have previously used PRAAT and now trying out Luscinia. Any other options out there? 
Relevant answer
Answer
Hi Adwait!
Have you checked warbleR ? This is a free R code developed by Marcelo Araya-Salas to streamline analysis of animal acoustic signals. 
Probably this is what you are looking for, although you must have some knowledge of R to use.
All the best!
  • asked a question related to Database Management
Question
4 answers
how to generate market basket dataset for about 50,00,000 transactions?
Relevant answer
Answer
Hello Saritha,
you can use data generator in the attached link to generate up to 10000 rows data in several file formats (CSV, Excel, SQL, JSON, HTML and XML).
You can construct the table fields, fill it with data and download it.
Regards 
  • asked a question related to Database Management
Question
5 answers
Ahoj everyone, I would be interested in a very simple application where somebody (both researchers, when working and the public, when waiting on a train or sitting on a public bench) could enter 3-4 ring colors of a color-ring marked bird plus either choose a location on a Google map or let the device save the GPS position together with the time stamp and the color code. So the app should just take these data and put it in a database. Any other functionalities like allowing the observer to add a behavioural category or similar would be nice but totally optional. Thanks for any hint whether this is existing somewhere or showing interest if you think the development of such a tool is in your field and scope of activity.
Relevant answer
Answer
Thank you Victoria and Rien for the hints to the "zooniverse" and the "vogel het uit!", but I think "animaltrack" is so far the platform which comes closest to what could be needed for such a mobile application. Thanks also to Giuseppe Masciopinto for the offer to develop such a tool. I will most probably come back to some of you in a later stage of my brain storming process!
  • asked a question related to Database Management
Question
1 answer
Specifically, I would like to be able to ascertain volume for structures such as the basal pons across the primate order.
Relevant answer
Answer
Hi David, The only database i am aware of is an old one from a series of papers by Barron et al. 
1. Baron, G., et al., Comparison of brain structure volumes in Insectivora and Primates. III. Main olfactory bulb (MOB). J. Hirnforschung, 1983. 24: p. 551-558.
2. Stephan, H., H.D. Frahm, and G. Baron, New and revised data on volumes of brain structures in insectivores and primates. Folia Primatologica, 1981. 35: p. 1-29.
These are not anything like as detailed as you might wish. I should think that any newer papers would have cited the above, so maybe there is something in the more recent.
  • asked a question related to Database Management
Question
10 answers
Does anyone have expertise or experience in creating a medical database, intended to collect/store clinical data? Which program would you recommend?
Or do you know of any source for more information? 
Many thanks. 
Relevant answer
Answer
Try Abstract Plus, developed by CDC USA, it is a cancer registry tool, stable and very helpful for single institution application.
  • asked a question related to Database Management
Question
1 answer
Greetings. I am looking for an on-line database where I can upload the MSMS peak list of identified proteins from gel based proteomics. Other than PRIDE, does anyone have any suggestions? This data need to be made available for publication. Thank you
Relevant answer
Answer
Dear Saiful Anuar Karsani,
The first requirement for database searching is a peak list; you cannot upload a raw data file. Raw data is converted into a peak list by a process called peak picking or peak detection. Often, the instrument data system takes care of this, and you can submit a Mascot search directly from the data system or save a peak list to a disk file for submission using the web browser search form. If not, or if you have a raw data file and no access to the data system, you’ll need to find a utility to convert it into a peak list. Peak lists are text files and come in various different formats. If you have a choice, MGF is recommended. Be careful with mzML, because this may contain either raw data or a peak list.
Regards, Shafagat
  • asked a question related to Database Management
Question
1 answer
how to use one hashing bloom filters to detect intrusion in a database management system
Relevant answer
Answer
try this URL, the logic for the algorithm can be found therein
  • asked a question related to Database Management
Question
6 answers
Should we be collecting data more intelligently rather than collecting all sorts of garbage and then apply "statistical hammer" to break it into smaller pieces? Why not use more intelligent statistical design to collect the data in first place?
Relevant answer
Answer
This discussion is very interesting! I may just provide an answer to your title question, as we proposed the following definition for Big Data in the recent paper attached: "Big Data represents the Information assets characterized by such a High Volume, Velocity and Variety to require specific Technology and Analytical Methods for its transformation into Value". In case you don't find it convincing, you will find a survey of other definitions inside the paper. Hope this helps, Marco
  • asked a question related to Database Management
Question
6 answers
I need to set up a database with the following requirements:
- easy handling (data entry, read out / statistics, changing the data entry interface and adding new parameters)
- in best case it should be web based --> the db file is on a server in our network and can be accessed (for data entry) by a web browser
What do you recommend? Filemaker or MS Access, or any other solutions?
Relevant answer
Answer
If you have to build everything from the beginning and if it's web based - Firstly you have to design database (E/R model for example), to define the tables, after that to define primary keys, foreign keys and references between the tables inside the database, to create tables and relations with SQL commands and in the end to start entering data and make requests after that.
  • asked a question related to Database Management
Question
1 answer
Commit protocol for DRTDBS
I am currently working to develop a new real time commit protocol for distributed real time database system.I need a help for simulation study of my work..
Relevant answer
Answer
Dear Researcher
please check this attachment.it might be helpful for you.
Regards
  • asked a question related to Database Management
Question
4 answers
if it is possible in which way we achieve it 
Relevant answer
Answer
Sure, you should explore:
- selecting an approppriate table type or storage engine: http://www.mysqltutorial.org/understand-mysql-table-types-innodb-myisam.aspx
  • asked a question related to Database Management
  • asked a question related to Database Management
Question
7 answers
if it is possible please suggest me methods to optimize the database connections of a particular website 
Relevant answer
Answer
It depends on the database itself. In software development particularly web application, you can't have optimised database straight away. Off course, the normal practices should be applied and the database should be optimised as it has to, but what I have learnt, some of the attributes of various tables are differently manipulated to that when it was designed. It is always a good idea to revisit most of the database as there is always a scope for improvement. 
Coming back to your question, more detail is required to get a precise response.
  • asked a question related to Database Management
Question
2 answers
Please provide the Name of the tool(s) and the link. Thank you in advance.
Relevant answer
  • asked a question related to Database Management
Question
2 answers
kd-trees introduced by J.L. Bentley, range trees, divided kd-trees etc. belong to a class of kd-tree variants for pointer machine model. However, Bkd-tree is a RAM model based kdtree,  and recent papers on both models are
a. Pointer machine model 
b. RAM Model
For comparisons, can kd-trees and its variants falling under pointer machine model of computation be combined with RAM model based kd trees?
Relevant answer
Answer
Thanks for the response.
  • asked a question related to Database Management
Question
2 answers
What are the applications of ERD and normalization?
Relevant answer
Answer
ERD(Entity-Relationship Diagram) is the conceptual model. It represents relationship between entities. ERD is converted into Relation model(tables) using concepts of normalization(rules to prepare tables to store data). Any application which requires data to be stored into database(tables) could be considered as application of ERD and Normalization. For ex., Hospital Management system, Financial Accounting System, Library Management System, Hotel Management System, Online systems to book place orders, book tickets, etc.
  • asked a question related to Database Management
Question
7 answers
Does anyone know about the which Database or Management Database System suitable to store OpenStack Ceilometer monitoring information for future processing? 
Thanks to all.
Relevant answer
Answer
Hi, we are currently following the approach to use a zabbix to collect monitoring information coming from ceilometer and other sources.  Depending on the scale of your infrastructure this could be a good approach since zabbix already allows to collect infrastructure related metrics in an easy way (via agent based approach or SNMP and many others) using its own templates and it is easily extendable to get information from ceilometer as well.
Alternatively you could converge all the information from your sources into elasticsearch (directly or using logstash) and use kibana for the analysis.
Cheers.
  • asked a question related to Database Management
Question
4 answers
I have calculated Transmit time of data packet as:packet length/data rate, now want to calculate receive time of data packet, because finally I have to calculate the delay, thanks in advance
Relevant answer
Answer
How you calculated transmit time, then I can exactly answer. Receive time would also have same things to be considered(hops processing, queuing, ready to transmit, recevie device acceptance time etc). I need to know what exactly you want? If you tell me how you have calculated transmit time then I can answer the exact thing.
  • asked a question related to Database Management
Question
1 answer
It is very important to keep a check on scalability and performance using APMs to solve big data problems. How important is the role of automated APMs (Application Performance Management) in solving Big Data problems?
Relevant answer
Answer
Support existing and new applications, voluminous data and storage performance
  • asked a question related to Database Management
Question
3 answers
comparison on the base
1. time require for retrieve from database a
2. database management of images for efficient search
etc
Relevant answer
Answer
Hi Ranjit,
If i'm not mistaken, you are asking, how can we evaluate and compare two different image retreival method. At this point, precision, recall, MAP and the F measure are matching with your needs.
  • asked a question related to Database Management
Question
4 answers
Hi all,
I need to use an index structure that require less complexity
first I opt for b-tree but I'm not sure that it's a good choise
in other hand a lot of index structure exits in lettrature
  • quadtree
  • kd tree
  • R tree
Thank you for help .
Relevant answer
Answer
Hi,
the choice of index type depends on your database type and use. If it is a relational database for an OLTP usage, a B-tree index is usually used. However, if your database is meant for business intelligence (modeled as a star schema) with read-intensive queries and a high selectivity ratio, other index types can me used such as bitmap, star join indexes, etc.
  • asked a question related to Database Management
Question
5 answers
NoSQL is highly valued for consistency, but is not strong enough to fit a set of ACID properties.
Relevant answer
Answer
I think that in the near future they will have to deal with some sort of consistency issues specially for critical applications. Most business databases (transactional) works with strong consistency but lack the high performance of NoSQL databases. Trading of performance with consistency is not a solution. some attempts have been done to add transactions to NoSQL databases such as http://www.vldb.org/pvldb/vol6/p1434-dey.pdf in VLDB2013. I believe it will come to a point in the future that there will be transactions (consistency) without losing the high performance of NoSQL databases.
  • asked a question related to Database Management
Question
3 answers
Database Management Systems for all purpose/type databases?
Relevant answer
Answer
The idea of separating the schema from data storage was proposed in an article in ACM Transactions on Database Systems in 2000 (Parsons and Wand 2000). In this instance-based data model, the schema can guide either the population of the database or as a mechanism to view (query) the data, or neither. However, instances can be accessed independent of any schema/classification. The model is compatible with column-based and graph databases.
Parsons, J. and Y. Wand (2000). Emancipating Instances from the Tyranny of Classes in Information Modeling. ACM Transactions on Database Systems, 25(2), 228-268.
  • asked a question related to Database Management
Question
8 answers
Are these problems the ones that are usually involved with new systems, where new skills are needed?
Relevant answer
Answer
look at our recent experimentation in a conversion of SQL to NoSQL and see the results. We suggest that some guidelines would be helpful because the current developers have years of SQL thinking
  • asked a question related to Database Management
Question
6 answers
Order_number         T_Calc_Time
2013040100086600 09:16:34
2013040100146880 09:19:13
2013040100406920 09:35:30
2013040100893120 10:13:09
2013040100086600 10:15:14
2013040101906200 12:07:59
2013040100146880 14:48:22
2013040103401460 15:06:34
2013040103407980 15:07:16
2013040103401460 15:09:11
2013040103407980 15:24:22
select Order_number, count(Order_number),T_Calc_Time,TIME_TO_SEC(TIMEDIFF(T_Calc_Time,T_Calc_Time)) from cash_order01 group by Order_number into outfile '/tmp/ordernumber1.csv';
I have to find out that I the order number come again then i have to calculate the difference between those times.
for example the order 2013040100086600 come at 09:16:34 and leave at 10:15:14 i have find the difference and print  for each order. If the order comes only once then we have say it is gone to next day.
can any one help me?
Relevant answer
Answer
Try this:
SELECT
  a.order_number,
  Min(a.t_calc_time) start,
  Max(a.t_calc_time) end,
  Timediff(MAX(a.t_calc_time), Min(a.t_calc_time)) time_difference
FROM
  ankit AS a
GROUP BY
  a.order_number;
It should work better and faster.
Try to limit the query for, say, a 100 records with LIMIT 0, 100 (at the end of the query) to see if the results are good?.
  • asked a question related to Database Management
Question
5 answers
Microsoft access is a software example for relational databases. I need more examples for relational databases. I need also some more examples for Object oriented databases and XML databases.
Relevant answer
Answer
Caché, ConceptBase, Db4o, GemStone/S, NeoDatis ODB, ObjectDatabase++, ObjectDB, Objectivity/DB, ObjectStore, ODABA, OpenAccess, OpenLink Virtuoso, Perst, Picolisp, siaqodb, Twig, Versant Object Database, WakandaDB, Zope Object Database.
  • asked a question related to Database Management
Question
5 answers
I have .csv file of size 15 GB. I used load data command in my sql to load the data to mysql table. It is skipping the records after 9868890. I read the maximum number of mysql table records are 5000000000. why I could not able to upload? I tried to increase number of rows by "alter table filename max_rows=200000000" then checked the status. It shows that max_row are that much but I could not able to insert not even one record more. please help me. Please dont ask me to split the file. I have 500 files like this and I have to upload. and My professor will not be happy if I simply break the files and do it.
Relevant answer
Answer
First, try to optimize your table's schema according to comments and suggested changes in this fiddle http://sqlfiddle.com/#!2/9bd7c/1/0 I've made for you.
Could it be that there is more than one record with the SAME combination of order_number, buy_sell_indicator and symbol in your CSV? I guess that the order number should be unique, but you should make sure that this is not the reason for failure.
Also, it says that there are warnings. Could you list some of the warnings in the SQL error log for me to see. Maybe there is some info that could help us find the problem.
  • asked a question related to Database Management
Question
4 answers
I am looking for some papers that address issues related to information architecture in the context of big data. Perhaps papers that broadly discuss the issue of large data sets on the web and how to deal with designing information architecture. Also, suggestions for research groups doing this kind of research would be appreciated. 
Relevant answer
Answer
If you are looking for big data storage model then this paper might be useful to you.
I know about one research group, its Big data working group at Cloud security alliance. but they are working on privacy and security issues.
  • asked a question related to Database Management
Question
3 answers
see above
Relevant answer
Answer
If the CouchDB is better than the other NoSQL stores it would be question of the application and the feature needed for a specific scenario.
CouchDB is designed for offline operation; it uses multi-master asynchronous replication. In CouchDB, multiple replicas can have their own copies of the same data, modify them, and then synchronize these changes at a later time. Comparison of different SQL implementations can be found on: http://publish.uwo.ca/~kgroling/NoSQL%20and%20NewSQL%20survey.pdf
  • asked a question related to Database Management
Question
4 answers
I have asked this question to collect information about documentation in general. And, to specifically know the procedures of documenting existing systems, especially the old systems that didn't have documentation.
Relevant answer
Answer
Use the MODEM Architeture Description Method and define the Configuration Baselines of your system elements in your Enterprise System Release. Do all this together with the Requirement Profiles for each element. Then you have defined your design of a Capability Package (CP)
/P
  • asked a question related to Database Management
Question
4 answers
All the good things we say about the GS database, such as open access, size, metrics, and algorithms, but when we use it for a scientometric study, we come to realize that the quality of the database is more important than its size. In searching an author query also retrieves citations of other authors with the same last name, but different initials for first and middle names. We cannot use this data until the unwanted records are removed. However, it can only be done done manually which may take hours if you have many authors
Having supervised the Cataloging Department for most part of my professional career, I would suggest that the GS should hire catalogers with the knowledge and experience in authority control and database maintenance. If they wish to outsource this job, there are many vendors available in the market.
I have proposed only one solution, What other things come to your minds. that you would like to suggest to improve the quality of the GS database?
Relevant answer
Answer
Those who can afford subscription-based citation databases, may not appreciate the value of GS for users in developing countries. The strength of GS is its retrieval rate of matching records between 92 and 100%. In our recent study, the hit rate was more than 98% The database has a good balance of low-impact and high-impact journals.For developing countries there is no other choice but to use Google,Scholar as they cannot afford to pay the high cost of Scopus or WoS. They have open access to GS.,the largest source database in the world and the possibility of  downloading approximately 20% articles in full-text..     
  • asked a question related to Database Management
Question
3 answers
I'm very new in the field of big data analysis and I strongly believe there is potential know-how that could be beneficial also in the field of corpus linguistics. Has anybody ever tried to merge corpus linguistics and big data methodologies?
Relevant answer
Answer
BigData:  The very question is whether there exists a schema (type) or not. If there is a schema we are very close to databases, except for the huge extension of "big". If not then information retrieval comes into play. with its indexing techniques.
H.Wedekind
  • asked a question related to Database Management
Question
5 answers
We know that data will deliver information which will form a knowledge. So, there exists database, information-base and knowledge-base. Will a database and a knowledge-base co-exist?
Relevant answer
Answer
I agree with Samir.. But KB has evolved..a 30 years ago KB is not what we define as a KB today
Then, Knowledge base , KB was  a DB+IE ( IE is inference engine ), IE convert  the  Data in to useful Knowledge..
Data models what we had 30 years ago, stored data as  "flat data", using hierarchical or Relational. They created Structured data..But most of the data we store today are Semi or unstructured data ( Facebook) , and hence data models that can manipulate unstructured data are necessary..
  • asked a question related to Database Management
Question
11 answers
Inmon, Kimball, Hefesto or another? I'm currently building a data warehouse to pave the way for data mining, the goal of this work is to improve the process of decision-making in education policy. This requires knowing what the best architecture is.
Relevant answer
Its depend on what do you need, no matter what the architecture, its go back to the purpose to build datawarehouse where we need to deliver database environment which can create best sql performance when access data from datawarehouse
you can create star schema or snowflake, top down or bottom up, its depend on how best performance your sql to access datawarehouse rather than from oltp.
  • asked a question related to Database Management
Question
2 answers
I need to know which Relational Data Base Management System manages derived data (Materialized view) in a Hierarchy? Please tell me the name of the tools which are common in practice such as Oracle, Sequel Server etc.
  • asked a question related to Database Management
Question
1 answer
I need different heuristics for query optimization.
Relevant answer
Answer
  • asked a question related to Database Management
Question
6 answers
What would be the best software for a large collection of varied specimens by different researchers for different uses?
Relevant answer
Answer
if you have SQL Language skills, you can use MySQL or Joomla for free
  • asked a question related to Database Management
Question
8 answers
I have proposed a special journal issue on 'Data Mining and Big Data' to IOSR (International Organization of Scientific Research) Iosr Journals. You can contribute your research/review papers (those who are interested). You can write your papers on the topics- Data mining, Databases, Management of Big data and other related fields.
Relevant answer
Answer
Its not exactly free, it charges an amount of rupees 400 for each selected paper for publication, but I have taken an initiative of giving the publication fee for all the selected papers myself (just for encouraging researchers).
  • asked a question related to Database Management
Question
7 answers
The database needs to handle large scale data (12 Mio entries at a time).
Relevant answer
MySQL is controlled by it self, as you know its a DBMS.
PHP is just used to retrieve data from MySQL database. Therefore, I think you should look for the MySQL DBMS optimization.
What type of data it is? What do you mean by processing 12M data at a time? If your database has 12M entries and you want to update and retrieve data, then you need to improve your searching algorithm in your PHP script. However, this will improve the performance of the program, not the MySQL database.
These are my 2 cents.
Good luck,
Sarves
  • asked a question related to Database Management
Question
9 answers
We are performing an international registry. Our database is online and each participating centre has its own user account to a private space in the database. This is like having many registries in parallel. For data quality checking we need to be able to track a record and communicate with the site if some data is out of range or strange. But I have been told that we are not allowed export the study ID into our research database. Is this true? What are best practice to link our data to the source data in anonymous way? One way is called cross referencing. Does anyone know a good reference for this?
Relevant answer
Answer
Hello, since the previous answers didn't mention it, I think you should look at anonymization techniques. The point is that even if you remove all obviously identifying fields, like names, unique IDs and so forth, it may still be possible to uniquely identify records by the association of particular fields. For example, Sweeney demonstrated about 10 years ago that 87% of the american population could be uniquely identified by an aggregation of Gender, Age and Zipcode where they lived.
Narayanan and Shmatikov also could deanonymize data published by netflix for research purposes, data that had been previously anonymized by removal of identifying fields.
If you are interested in this, I would suggest you look at these specific terms: k-anonymization, l-diversity, t-closeness. These all apply to manipulations of database records. You lose in accuracy of the data, but you gain in privacy. If you want to keep the data as exact as possible but only for aggregate queries, then you might be better served by other techniques like data perturbation and permutation of sensitive data.
  • asked a question related to Database Management
Question
5 answers
I'm trying to find out some information/journal articles on the impacts of implementing a database management system or implementing databases into organisations and how they can help. If someone could guide me in the right direction please that would be great.
Relevant answer
Answer
Perhaps you are looking for information regarding Return on Investment (ROI) to guide a proposal to develop or implement DBMS? ROI estimates should include both tangible financial costs and returns and intangible costs and benefits. As mentioned by Venkateswara above, this really does depend on the sector you are in and the type of system and purpose of the system. ROI seems to be much more easily measured in commercial and financial sectors, but much more difficult to evaluate eg in the health sector, where there is still debate about the benefits of HIT. The benefits derived from a database system (especially if this is not also part of an operational/transactional system) will also depend on the ability to use the stored data afterwards, so benefit can be confounded by poor management and use after implementation. Try looking up ROI for information systems in the relevant business area.
  • asked a question related to Database Management
Question
3 answers
I would like to know if there's a way to resize it up to a desired size, or if it is decided by MSSQLS.
In any case, how to do that without compromising the whole database?
Relevant answer
Answer
In database file properties you can set the size of the file after shrinking.
  • asked a question related to Database Management
Question
13 answers
Each group of synonyms (English) is linked to a bibliographic reference. The database (Bibliomac for Macintosh only) is driven by a software called 4D. Unfortunately this database is no longer developed. It's a pity because the possibility to link a group of synonyms with bibliographic references is uncommon and powerful for data mining. Reference Manager can do something similar but I don't know how to transfer this list from bibliomac to it because I'm not a computerist? Could somebody help me? I can send the database and the software (unfortunately in French).
Relevant answer
Answer
Monika, thank you for your kind remarks.Johannes is right . My main purpose is to vonvert an old fashioned and obsolete database ( Bibliomac) in a more modern but in keeping the main functoinnality of bibliomac: the ability ro link a bibliographic reference to a group of synonyms. End note is not able to do it. Would be Zotero ?
UMLS and MESH ontology is very good to manage synonyms of simplesemantic entities ( for example: adrenoceptors/adrenergic receptors) but is less efficient for complex query : for example VEGF expression /production/ secretion by cancers/tumors or Effects of thyroid hormones on heart. The use of the boolean operator AND does not work always because it catch up the different terms without respecting the hierarchy of the sentence (of, in,by,) and broadening too much the research ( for example cancer, cancer cells, cancer cell lines). The use of the entire sentence and its synonyms avoid this .It's the main reason , I think, my list could be useful if the link with the references is kept on.The second reason is that these descriptors have been made after a reading of the full text of the article and can't be deduced from the title , the abstract or the author's key words.Ideally, It could be the first step to build a knowledge database with the help of a network of biomedical researchers ( following the model of the protein-protein interaction of the Van Buren Lab). What do you think about this possibility ?
  • asked a question related to Database Management
Question
3 answers
These could include the number of joins, number of records in the tables, number of locations where data is stored, number of joins, join order etc.
Opinions on non-SQL databases may also be discussed.
Relevant answer
Answer
One thing I usually consider is if DBMS installed and provided in the Cloud being analysed for possible in-productivity use supports triggers, partitioning, complex indexing by use of regular expression models etc.
As for number of joins, records, tables etc. it has more to do with the quality of the database design and normalization level than with the cloud solution that provides the database system.
I too would like to hear of some key-value type DBs in he cloud if someone after me comments and has experience with them.
  • asked a question related to Database Management
Question
7 answers
I want to compare diff database solution with respect to distributed database features availability.
Relevant answer
Answer
Can you write up some details as to how and under which scenarios you want to use the database solution?......Generally, we have Oracle 11g and IBM DB2, but these are for really big enterprises......An alternative to MySQL is PostgreSQL.....Also there are quite a few different types of storage engines in MySQL suited for different needs......One other dimension is databases for Mobile Devices where you have SQLite. So, its basically more about the usage.....I can help you out on that