Science topic

Information Retrieval - Science topic

Organized activities related to the storage, location, search, and retrieval of information.
Questions related to Information Retrieval
  • asked a question related to Information Retrieval
Question
2 answers
To handle second order factor, I used the latent variable scores calculated by smartpls. After attaching that data file with my path, i tried to create groups on the basis of a categorical variable. But it simply doesn't work. It doesn't give an error either. Has anyone faced the similar problem ?
Relevant answer
Answer
Thanks. Please contact support@smartpls.com --> The SmartPLS team will surely be able to help you solve this issue.
  • asked a question related to Information Retrieval
Question
2 answers
I have done Rietveld refinement on PXRD but i do not know where to find the information of Final R indices[I>2sigma(I)] , R indicies (all data) , Extinction coefficient, Largest diff. peak and hole.
Relevant answer
Answer
Actually, it depends on the software you have used for the Rietveld refinement. As a JANA2006 user, I would try to export a CIF file and open it with any text editor (e.g., Notepad) afterwards. Such information is usually automatically written into the CIF file. If not, you may try to read the log files for the refinement procedure (for instance, in JANA2006, full details on the refinement procedure are given in the respective *.ref file)
  • asked a question related to Information Retrieval
Question
2 answers
Dear researchers,
Please help me with this issue.
I used Delvetool.com to analyse my qualitative data after I found my edited transcripts were lost in Nvivo and no cause could be found for it. I also felt that Nvivo is quite complicated to use, while Delve is much more straightforward. However, the ethics committee at my university had never heard of Delve before and raised a concern to me about the data storage in Delve. Delve uses 3rd party processors, which means that the data uploaded in Delve is stored in a number of external servers. Delve developers argued that they have listed all 3rd party processors on their website and they are all well-known and trusted (https://delvetool.com/subprocessors).
I have searched and found a few publicated papers that used Delve for thematic analysis. It is really hard for me to go back and use Nvivo now. Could you please advise me if there would be serious ethics concerns to uploading data into Delve and using it for thematic analysis?
Thanks a lot!
Sincerely,
Siying
Relevant answer
Answer
It sounds like the basic issue is the security of data, and whether you used aa thematic analysis etc. is not really relevant.
As far as I can tell, Delve its a "cloud-based" analysis system, so a likely comparison is to Dedoose, which was the first program to use this technology. I read over both of their privacy policies, and one thing that I noticed is that Dedoose explicitly states that it stores your all information in an encrypted format, while I did not find any such statement from Delve.
So, I think you should contact Delve, tell them your ethics committee has a concern about data security, and in particular ask whether they store your data in an encrypted format.
  • asked a question related to Information Retrieval
Question
2 answers
Hi All,
Question: How do I get Mach No contour for CFD Post-processing?
Description:
  • Solution computed using HPC
  • Cannot open the file back in fluent with Case & Data to export Mach No variable.
  • However, with the data file, I can open it in the post but no Mach number is available.
What could I try?
Any help is appreciated.
Thank You,
Alin
Relevant answer
Answer
Hi
please say how solve it. I choose the Mach number from data quantity files in fluent but it does not exist in cfd post
Thaank you Alin Sonny
  • asked a question related to Information Retrieval
Question
3 answers
I was reviewing a paper for a journal, and the authors stated that they randomly selected company reports from several databases to give them a sample size of 600 firms.
I was of the opinion that “sampling” is inherently associated with primary data collection or am I mistaken?
My suggestion would be that the authors state that they selected 600 “cases” for their analysis from the databases (using whatever criteria) and not make reference to the word “sampling”…?
Relevant answer
Answer
The term "sampling" can indeed be used when utilizing secondary data sources for a study. Sampling refers to the process of selecting a subset of individuals, cases, or data points from a larger population or dataset for analysis. This process can apply to both primary data collection and secondary data analysis.
In the context of the paper you reviewed, where the authors selected company reports from databases as their sample, it is appropriate to use the term "sampling" to describe their process. Even though they did not collect the data themselves through primary research, they still sampled a subset of data from the larger population of available company reports.
It's important to note that the term "sampling" can be used in different research contexts, including both primary and secondary data analysis. However, if the authors prefer to be more precise and avoid any potential confusion, they could explicitly state that they selected 600 "cases" or "data points" for their analysis from the databases instead of using the term "sampling." Ultimately, it would depend on the specific terminology and conventions used in the field and the preferences of the authors and the journal they are submitting to.
As a reviewer, you can certainly provide your suggestion to the authors regarding the use of terminology, offering the option of stating that they selected 600 "cases" rather than using the term "sampling." However, it would be a judgment call for the authors to decide whether to adopt that suggestion based on their understanding of the field's conventions and the specific context of their study.
  • asked a question related to Information Retrieval
Question
3 answers
Laptop / pc recommendations for efficient writing, endnote, spss, data storage?
Relevant answer
Answer
Processor: intel core 5 or Ryzen 5 or higher
RAM: 8 GB or more
Storage: 256/512 SSD and 1 or 2 TB HDD
SSD can be used for loading OS and programmes. HDD is the safer option for data storage.
Screen size: 14" or 15.6"
14" is more compact, but 15.6 provides enough screen space. Choose the size according to your convenience.
You can buy an additional screen (22" or higher) to connect externally and keep it in your office or home. That will be a great option when you are working with huge data, charts etc.
New laptops have enough features, including USB ports, card reader, backlit keyboard etc.
Graphics cards: A 2GB graphics card would be an additional option, and it will definitely help to handle graphics data if you need to work with 3D graphics and videos. The more 'GB' (4GB, 6 GB), the better the graphics performance. But the size of the laptop will be increased. If you need a slim laptop with greater graphics, you may need to pay a higher amount.
I think for your need, 2 GB graphics would be enough.
I recommend buying an external HDD (not SSD) for data backup. I think 2 TB would be enough. (WD or Seagate are good, ensure password protection feature)
The laptop brand is a personal choice.
If you are comfortable with Apple, go for Macbook. But in my personal experience, the device is so limiting yet efficient!!
  • asked a question related to Information Retrieval
Question
1 answer
clc;clear;
% Create an empty array to store the electron density data
ne_array = [];
% Loop through the files in the folder
folder_path = 'D:\ionPrf_prov1_2020_002'; % replace with the path to your folder
file_list = dir(fullfile(folder_path, '*.0001_nc')); % replace '.0001_nc' with the file extension of your data files
for i = 1:numel(file_list)
% Read the electron density data from the file
filepath = fullfile(folder_path, file_list(i).name);
fid = fopen(filepath, 'r');
while ~feof(fid)
line = fgetl(fid);
if startsWith(line, 'ELEC_dens') % look for the line starting with 'Ne'
ne_data = strsplit(line);
ne_data = str2double(ne_data(2:end)); % extract the Ne data as an array of doubles
ne_array = [ne_array; ne_data]; % add the Ne data to the array
end
end
fclose(fid);
end
% Save the electron density data to a text file
output_filename = 'ne_data.txt';
dlmwrite(output_filename, ne_array, 'delimiter', '\t', 'precision', '%.3e');
Relevant answer
Answer
The code you provided seems mostly correct. However, there is one potential issue that could cause an error. In the line `ne_data = str2double(ne_data(2:end));`, the indexing operation `ne_data(2:end)` assumes that there is at least one value in `ne_data`. If the line starting with `'ELEC_dens'` is not found in a file, `ne_data` will be an empty cell array, and indexing it will result in an error.
To resolve this issue, you can add a condition to check if `ne_data` is not empty before performing the indexing operation. Here's the modified code:
###Matlab start
clc;
clear;
% Create an empty array to store the electron density data
ne_array = [];
% Loop through the files in the folder
folder_path = 'D:\ionPrf_prov1_2020_002'; % replace with the path to your folder
file_list = dir(fullfile(folder_path, '*.0001_nc')); % replace '.0001_nc' with the file extension of your data files
for i = 1:numel(file_list)
% Read the electron density data from the file
filepath = fullfile(folder_path, file_list(i).name);
fid = fopen(filepath, 'r');
while ~feof(fid)
line = fgetl(fid);
if startsWith(line, 'ELEC_dens') % look for the line starting with 'Ne'
ne_data = strsplit(line);
if numel(ne_data) > 1
ne_data = str2double(ne_data(2:end)); % extract the Ne data as an array of doubles
ne_array = [ne_array; ne_data]; % add the Ne data to the array
end
end
end
fclose(fid);
end
% Save the electron density data to a text file
output_filename = 'ne_data.txt';
dlmwrite(output_filename, ne_array, 'delimiter', '\t', 'precision', '%.3e');
##code ends
With this modification, the code will only perform the indexing operation and add the data to `ne_array` if `ne_data` has more than one element, ensuring that an empty line or a line without the desired data doesn't cause an error.
  • asked a question related to Information Retrieval
Question
2 answers
I frequently make use of LigParGen website to generate a molecule OPLS-AA data file for Lammps by uploading a .pdb file.
Generally it does not take a long time to get the lammps file.
Now after a long time following the .pdb file upload I am receivingng an out -of-time error message.
Anybody is experiencing the same problem?
Thanks
Paolo Botto
Relevant answer
Answer
Not at all.
  • asked a question related to Information Retrieval
Question
4 answers
thesis or scientific research
Relevant answer
Answer
Several proposals exist that involve using smart contracts to collect and process data from IoT devices. One proposal outlined in a paper by Hang and Kim involves a blockchain platform for securing IoT sensing data integrity . The platform provides smart contracts for registering sensors and collecting data from them . The proposed solution aims to automate the process of collecting and verifying IoT sensing data, ultimately improving the overall quality and accuracy of the data.
Another proposal involves using smart contracts to automate agriculture data processing , as outlined in a paper by an unknown author. The proposal uses smart contracts to process and automate data collected by agricultural sensors . This allows for improved accuracy and quality of the data, as well as increased efficiency in data processing and analysis.
In both proposals, the use of smart contracts enables automation and efficiency in collecting and processing data from IoT devices. By using smart contracts for these processes, the need for intermediaries and manual verification is reduced, resulting in improved accuracy and efficiency.
  • asked a question related to Information Retrieval
Question
1 answer
Is Secondary Research valid for the publications as its data sources from the primary research?
Relevant answer
Answer
Secondary research is valid for publications as its data sources can come from primary research, as well as other sources such as academic journals, public libraries, open data, government publications, and databases. When conducting secondary research, it is important to be thorough and ensure that the sources being used are credible and up-to-date. While primary research involves collecting data directly from the source, secondary research involves analyzing and synthesizing existing data from various sources. Both primary and secondary research can be valuable in generating insights and informing decision-making processes.
  • asked a question related to Information Retrieval
Question
3 answers
Basically I want to generate a mapping between uris(rdf,rdfs,owl) and NL keywords for a distributed meta-meta information system to assist SPARQL query construction from natural language query using controlled vocabulary like wordnet. For that I have to crawl linked open data to get uris describing entities and it's best matched keyword (e.g Rdf: label).
Relevant answer
Answer
Some tools used to extract only some properties from URIs while crawling on the fly over a web of data are RDF2Go, Sesame, Jena, OpenRDF, Sindice and RDFLib. These tools are used to extract data from RDF documents and RDF stores.
  • asked a question related to Information Retrieval
Question
4 answers
Data Collection Methods Via Secondary Sources.
Relevant answer
Answer
Secondary data methods refer to the techniques used to collect and analyze data that has already been collected by someone else for a different purpose. These methods are commonly employed in research and analysis to complement primary data collection efforts or to answer research questions that can be addressed using existing data.
It is important to note that while secondary data sources offer convenience and cost savings, they may have limitations such as data quality, relevance to the research question, and potential biases. Researchers need to critically evaluate and validate the secondary data sources before using them in their analysis.
  • asked a question related to Information Retrieval
Question
3 answers
Block chain
Relevant answer
Answer
Since the blockchain offers pseudonymity to the individuals, their activity on the blockchain can be linked back to them if privacy compromises. Access control, Security & encryption mechanisms and identity management like self sovereign identity management can help in addressing the privacy issues.
  • asked a question related to Information Retrieval
Question
2 answers
Dear All,
I'm looking for some Classified Partial Discharge datasets (Corona Discharge, Arcing Discharge, Surface Discharge, Internal Discharge) for pattern recognition study. I'd really appreciate if you could kindly mention some data sources from where I'd download the datasets.
Best Regards
Relevant answer
Answer
I also want to partial discharge datasets.
  • asked a question related to Information Retrieval
Question
6 answers
I am wondering if Google scholar has preset preferences with respect to location, race, etc. when it comes to displaying Search results. I noticed that when some of my scholarly articles are searched for on the platform using THE EXACT ARTICLE TITLE, such does not come up in the results whereas closely-related AND dissimilar articles by authors from some other parts of the globe does. Yet, my articles are indexed on Google Scholar and visible on my profile. Some only come up when BOTH my name and article title are searched for together on the platform. What could be responsible? I will appreciate any insight, thanks!
Relevant answer
Answer
If you don't put quotation marks around exact title searches, Google Scholar will find all articles with the same words, regardless of word order. Quotation marks ensure Scholar searches your title as a locked-in phrase. I see no signs of Google Scholar having a search bias.
  • asked a question related to Information Retrieval
Question
4 answers
Hi there,
I am seeking a reliable data source or database that can provide current and past affiliations for authors of scientific publications.
While I am familiar with platforms such as Scopus and Web of Science, I am unable to use them due to the requirement of an institutional subscription or the need to search for authors individually. Ideally, I am looking for a database where I can extract information on all authors affiliated with a specific institution.
My research is focused on computer science, so a data source that specializes in this field would be perfect, but I'm open to any suggestions you might have. Thanks!
Relevant answer
Answer
Big thanks to both of you Anne-Katharina Weilenmann and Alexander Schniedermann . Will look into these options shortly ^^
  • asked a question related to Information Retrieval
  • asked a question related to Information Retrieval
Question
2 answers
Hey !
I have been looking for the limitations for consurf server. As I have a data file containing 95k sequences but the job in consurf is still processing from last 3 week. So I wanna know is there any limitation for data file size (No of sequences) in ConSurf.
Relevant answer
  • asked a question related to Information Retrieval
Question
2 answers
If you have access to them please share with me. Thanks
Relevant answer
Answer
Thank you very much!
  • asked a question related to Information Retrieval
Question
4 answers
Given the issue of copyright of source publications, does the use of ChatGPT for the automated creation of new texts used in specific practical and commercial applications raise specific ethical issues?
Given the issue of copyright of source publications, the use of ChatGPT for the automated creation of new texts for certain practical, business and commercial uses may raise specific ethical issues if ChatGPT's creation of new texts uses certain source publications downloaded from the Internet and is not adequately acknowledged in the source footnotes. In such a situation, copyright may not be respected, which is a serious drawback, a breach of current standards for the use of source publications, and may seriously limit the practical use of new texts created in this way. Well, as a standard, ChatGPT does not provide a list of data sources in the responses. It is possible to ask for these data sources additionally and then it provides them, but there is no certainty that it provides them all. Sometimes, for general concepts, it lists sources such as textbooks, industry monographs, etc., and adds a statement that, in addition to these sources, the ChatGPT has also used so-called 'own sources', i.e. sources drawn from a knowledge base of several tens of terabytes obtained in 2021 from the Internet and contextually selected in relation to the question asked and, possibly, the preceding question's description of the context. The ethical issues related to the use of ChatGPT for the creation of texts used for specific practical and profit-making applications by freelancers, where a certain amount of creative work is required, are determined, inter alia, by the attitude of the person, company, institution or other entity using this tool to the data available on the Internet. Well, not all persons and entities using Internet resources treat the issues of openness of data and information provided on the Internet in the same way. There may be different approaches to the issues of demonstrating data sources, using them, respecting copyright. As a standard, i.e. according to the applicable legal and ethical regulations, even for data published under the open access formula, when using and writing texts, the sources of data, sources of inspiration, etc. must be indicated in the form of footnotes with information allowing to identify the specific source of the data. If this important issue is omitted and the sources of data, information, inspiration, specific statements, theses, explanations of concepts, etc. are not shown in a text that should be a new creative text, then serious drawbacks may arise both in terms of respecting copyright and the possibility of developing research in a given field, in a given topic, and in terms of verifying the veracity of specific information that ChatGPT originally took from the Internet (as of 2021 and according to a specific part and not the entirety of the data available on the Internet). If these issues are not met and the issue of copyright is treated with discretion, certain ethical considerations arise, i.e. the failure to comply with certain ethical principles. Besides, the issue of precise demonstration of data sources is also important for being able to verify the veracity of the data and information contained in the ChatGPT-generated answer to a question, i.e. the automatically generated specific text. The importance of this issue is directly proportional to the scale of errors and fictitious information generated by the ChatGPT of non-existent "facts" appearing in the texts generated by this system that are answers to the question asked. And, unfortunately, the scale of these errors and non-existent "facts" generated by ChatGPT, fictitious "data and information" created within the "free creation" of this system is not small.
In view of the above, I address the following question to the esteemed community of scientists and researchers:
Considering the issue of copyright of source publications, does the use of ChatGPT for the automated creation of new texts used in specific practical and profit-making applications generate specific ethical problems?
What do you think about it?
What is your opinion on this subject?
Please respond,
I invite you all to discuss,
Thank you very much,
Warm regards,
Dariusz Prokopowicz
Relevant answer
Answer
All the hype surrounding ChatGPT suggests that you can just ask it whether copyright would be violated if the content were distributed or what attribution would be required to avoid copyright violation. Maybe the problem can even be avoided with a suitable construal of harm in Asimov's First Law of Robots....
  • asked a question related to Information Retrieval
Question
3 answers
Hi everyone, I'm trying to convert WRF data to ARL format to run trajectories, using the Hysplit software. But, I have the next error:
child killed: segmentation violation
child killed: segmentation violation
while executing
"exec $exec_dir/arw2arl.exe $data_file >STDOUT"
(procedure "xtrct_data" line 37)
invoked from within
"xtrct_data $gdir $gbase"
invoked from within
".arwarch.bot.save invoke"
("uplevel" body line 1)
invoked from within
"uplevel #0 [list $w invoke]"
(procedure "tk::ButtonUp" line 24)
invoked from within
"tk::ButtonUp .arwarch.bot.save"
(command bound to event)
Previosuly, i converted the wrf data successfully!
Thanks
Relevant answer
Answer
Andrea Ximena Miranda Corzo The error message "child killed: segmentation violation" indicates that there is an issue with the program you are executing, not necessarily with the data itself. Here are some troubleshooting and resolution actions you might try:
1. Examine the input data: Check that the WRF data you're attempting to convert is in the right format and that the data files are free of mistakes. Before attempting to convert the complete dataset, you should convert a small portion of the data to test whether it works.
2. Examine the HYSPLIT installation: Ascertain that HYSPLIT is correctly installed on your machine, as well as all essential libraries and dependencies. You might try reinstalling HYSPLIT.
3. Examine the conversion program: The error message you're seeing implies that the conversion tool you're using ("arw2arl.exe") is to blame. Read the software's documentation to ensure you're using it appropriately, and consider using an alternative program or technique to convert the data.
4. Get assistance from the HYSPLIT user community: HYSPLIT offers a user community where you may seek assistance from other users who may have experienced similar problems. You can get advice from experienced HYSPLIT users by posting your query in the HYSPLIT forum (https://www.ready.noaa.gov/hysplitusersguide/SiteSetup-HYSPLIT forum.php).
  • asked a question related to Information Retrieval
Question
2 answers
Dear Professors/Researchers,
I would like to learn ANN process (For IC Engine related). I have seen many videos related to ANN in YouTube. Also, I read many articles related to ANN. The articles provide only results. But I can't understand how the input/output parameters are assigned/simulated in the ANN (For the IC Engine related). If possible, Kindly provide and guide with any reference data file or videos. Thanks in advance.
Relevant answer
Answer
Artificial neural networks (ANNs) are a type of machine learning model that is inspired by the structure and function of the human brain. ANNs consist of interconnected nodes or neurons that process information, and they can be trained on input/output data to learn complex patterns and relationships.The input data for an ANN can come from a variety of sources, depending on the specific application. In the case of an engine, the input data could include parameters such as temperature, pressure, fuel flow rate, and other sensor readings that provide information about the engine's operating conditions. This data can be used to predict the engine's performance, detect faults or anomalies, and optimize its operation.The output data from an ANN can also vary depending on the application. In the case of an engine, the output data might include predictions of engine performance, such as power output, fuel efficiency, and emissions. It could also include alerts for potential faults or anomalies that could indicate a problem with the engine.The specific input/output data used in an ANN will depend on the specific application and the goals of the model. However, in general, ANNs are flexible and can be adapted to a wide range of input/output data types and formats, making them a powerful tool for a variety of applications.
  • asked a question related to Information Retrieval
Question
5 answers
The data must be present as time series.
Relevant answer
Answer
Real-time data sources for power systems and electrical machines can include sensors that measure parameters such as voltage, current, temperature, and vibration. These sensors can be installed directly on the machines or on the surrounding equipment, and can send data to monitoring and control systems in real-time. Other sources of data can include digital meters, protective relays, and SCADA (Supervisory Control and Data Acquisition) systems. All of these sources can provide important information for monitoring the health and performance of power systems and electrical machines.
  • asked a question related to Information Retrieval
Question
17 answers
These datasets will be used for data classification and predicting new information
Relevant answer
Answer
The Aidos-x system is an excellent intelligent system for diagnosing and classifying human diseases. The methods of using the Aidos-x system for diagnosing human diseases are disclosed in lectures with sound "Using automated system-cognitive analysis for the classification of human organ tumors", "Intelligent system for diagnosing early stages of chronic kidney disease", which can be downloaded right now from the website https ://www.patreon.com/user?u=87599532 Creator's title: «Lectures on Electronic Medicine». After subscribing to this site, you will receive databases for medical research to identify the diseases that you will read about in lectures. The acquired skills of working in the Aidos-x system will allow you to apply for grants to carry out scientific research in the field of medicine.
To subscribe to the site https://www.patreon.com/user?u=87599532 you do not need to go to the bank, but you can do it using the Pay Pal system. Send a transfer in your currencies, and the transfer to dollars will be done automatically. After subscribing on the site, you will receive the Aidos-x system with an English user interface for free.
Thank you.
  • asked a question related to Information Retrieval
Question
3 answers
When I click on import data, I choose the SPSS data file (.sav) option and then select the file. In the next step, I click on the save button for the program to save the file as LISREL data (.LSF). Then, the error message comes up stating that ''Error importing, ... [location and name of the file] does not exist." I am repeatedly receiving this error. I changed the name of the file and reduce the variables in the dataset to minify the size of the file but they have not helped me so far. Is there anything I missed and how can I work out this problem?
Thank you in advance,
Relevant answer
Answer
Thank you, Dear Miller and Dear Al Zaim,
I synchronize my files with One Drive, once I copied the SPSS file to my hard disk that is not synchronized, the problem is figured out. However, I am going to try your suggestions.
  • asked a question related to Information Retrieval
Question
1 answer
I am looking at identified candidate genes for isolated orofacial clefts identified in various populations to see if the associations are replicated in a UK population. (Along with environmental exposures).
I will be using a secondary data source only, and I am aware of papers that explain what analysis is needed for primary data but I haven't seen if this differs for the use of secondary data?
Relevant answer
Answer
Georgia Awcock Case-control Secondary data genetic association studies often employ a variety of statistical approaches to examine the data and assess the strength of the relationship between genetic variations and the illness of interest. Among the most frequent approaches are:
1. Logistic regression: This approach is used to calculate the odds ratio (OR) and 95% confidence interval (CI) for a genetic variant's connection with illness. This may be used to analyze both the main impact of the genetic variation and the interactions between the genetic variant and environmental exposures.
2. The Cochran-Mantel-Haenszel (CMH) test is a statistical test used to examine the relationship between a genetic variation and illness while controlling for other factors such as population stratification.
3. Haplotype analysis: This approach is used to examine the inheritance patterns of numerous genetic variations in linkage disequilibrium. It can be used to identify particular haplotypes linked to the condition.
4. Multivariate analysis can be used to investigate the combined influence of several genetic variations and environmental exposures on the illness of interest.
It should be noted that when utilizing secondary data, it is critical to ensure that the data is of high quality and that the genetic variations and environmental exposures were quantified in the same way as in the main study. Furthermore, before beginning the study, it is critical to have a well-defined research topic, a sample size calculation, and a pre-specified plan of analysis.
  • asked a question related to Information Retrieval
Question
3 answers
I have successfully applied non-linear langmuir and freundlich isotherms in origin. However, the fit does not converge in case of Sips model. I am unable to solve the issue behind this. I would be really thankful for help in this regard.
PS: I have attached the data file here.
Relevant answer
Answer
Your data opens empty in the Origin workbook. Might be able to help if you can share an Excel sheet instead.
  • asked a question related to Information Retrieval
Question
2 answers
I need company specific data on company variables for past ten years for my reaserch
Relevant answer
Answer
Abderrahmane Boudribila Thankyou for your answer. Highly appreciate it. However, the SC link is not functional. Kindly provide me a direct link if its available. Also the BM only provides Financial Reports but not the data base like Investing.com provides.
  • asked a question related to Information Retrieval
Question
4 answers
What are some of the accessible yet trustworthy sources of secondary data in the field of logistics and supply chain management?
Relevant answer
Answer
There are several sources of secondary data in the field of logistics and supply chain management that are both accessible and trustworthy. Some examples include:
  1. Government agencies: Many government agencies, such as the United States Census Bureau and the United States Department of Transportation, collect and publish data related to logistics and supply chain management.
  2. Professional organizations: Professional organizations, such as the Council of Supply Chain Management Professionals (CSCMP) and the Institute for Supply Management (ISM), often publish reports and data on logistics and supply chain management trends and benchmarks.
  3. Industry associations: Industry associations, such as the National Retail Federation (NRF) and the National Association of Manufacturers (NAM), often publish data and reports on logistics and supply chain management issues that are specific to their industries.
  4. Market research firms: Market research firms, such as Gartner and Forrester, often publish reports and data on logistics and supply chain management trends and issues.
  5. Academic journals: Many academic journals, such as the Journal of Business Logistics and the Journal of Management, publish research on logistics and supply chain management topics.
It's important to carefully evaluate the credibility and reliability of any secondary data sources you use to ensure that the data is accurate and relevant to your research needs.
  • asked a question related to Information Retrieval
Question
3 answers
can any one help me to remove ransom virus (mppn and mbtf). all my data file are not working and the hacker asked me to pay 1000$ to return the files.
Relevant answer
Answer
Disconnect from the internet
Run a ransomware removal tool
Restore from a backup
  • asked a question related to Information Retrieval
Question
1 answer
task # 0
from pw_readfile : error # 1
error opening xml data file
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
task # 2
from pw_readfile : error # 1
error opening xml data file
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
task # 3
from pw_readfile : error # 1
error opening xml data file
Relevant answer
Answer
There could be several reasons why you are experiencing an error when attempting to open an XML data file using the Burai software. Here are a few potential causes and possible solutions:
  • The file path is incorrect: Make sure that you are specifying the correct file path for the XML data file. Check for typos or other mistakes in the file name or path.
  • The file is damaged or corrupted: If the file has been damaged or corrupted, it may not be able to be opened by the Burai software. Try opening the file in a different program (e.g., a text editor) to see if it is readable. If the file is damaged, you may need to try to repair it or obtain a new copy.
  • The file is not in the expected format: Make sure that the file is an XML data file and that it is in the correct format for the Burai software. If the file is not in the expected format, the software may not be able to read it.
  • There is an issue with the Burai software itself: If you are experiencing the error when trying to open multiple files or if you have tried the above solutions and the error persists, there may be an issue with the Burai software itself. Try restarting the software or reinstalling it to see if this resolves the issue.
I hope this information is helpful. If you continue to experience problems with the error, it may be helpful to consult the documentation or support resources for the Burai software or to seek additional assistance.
  • asked a question related to Information Retrieval
Question
4 answers
Can anyone suggest me a free plataform or API to retrieve data from sensors (in real time if possible)?
The data type can be about weather, agriculture or similar.
Relevant answer
In fact I only hope to retrieve real data (through an API, for example) to use data interpolation techniques.
I believe that there must be some API that allows me to retrieve data from sensors as close as possible to real time.
  • asked a question related to Information Retrieval
Question
3 answers
Hello professors
Hope you're doing well.
Is is possible to download data from NASA website and use it in research papers written for international journals. If possible then is the data source is authentic to use for research work?
For your information the website of NASA is given below from where I want to download and use data....
Please have a look and give your kind suggestions regarding the data authenticity.....
Thanks in advance
Regards
Sami
Relevant answer
Answer
Sami Ullah Khan The NASA Open Data Portal is NASA's clearinghouse site for publicly available open data, with access to tens of thousands of datasets.
NASA's Earth Observing System Data and Information System (EOSDIS) is built as a distributed system, with significant facilities situated around the United States at NASA's Distributed Active Archive Centers (DAACs).
  • asked a question related to Information Retrieval
Question
1 answer
Hi,
Did anybody used GOplot in R, I have the code running with EC.rda data file supplied with the code but I can not run it with my data (using .csv file). Can someone share their raw data file for me to have a look or can suggest how can I make it work.
Thank you.
Relevant answer
Answer
Yours faithfully.
Mónica.
  • asked a question related to Information Retrieval
Question
2 answers
Hello,
I would like to know what statistical analytical tool/model can be used on a dataset where price comparison data of daily necessities are recorded weekly from primary and secondary data sources.
The data is collected on a 30-week time frame and collected every Wednesday.
I have considered using trend analysis and time series, but I am looking for any analytical tool/model that goes beyond these two.
If you would kindly let me know what kind of interpretation or results I am likely to get with the analytical tool/model in question, that would also be great.
Relevant answer
Answer
Statistical Process Control Charts (aka Shewart Charts)
2-sigma (run-charts, p-charts, X-bar R charts, Moving average charts) are all excellent methods for visualizing process variation over time and identifying "special cause" variation patterns.
For predictive modeling: CHAID (chi-square-automatic-interaction-detection), BLR (binary logistic regression) can provide ranked hierarchy among the system of independently predictive variables in the model. CHAID shows the n-dimensional higher order interactions in a ranked hierarchy, BLR shows the Odds Ratios (compared to 50/50 chance), the "times more likely for the Dependent Variable state, if the independent factor is true". Both methods produce model probability "c-statistics", that can be utilized to calculate sensitivity, specificity, positive predictive value and negative predictive value as well as to plot and calculate the area under the receiver operating characteristic (AUROC) curve also known as "Signal to Noise Ratio" for the model. Using both models, the ROC Curve for each model can be displayed in the same plot area for simultaneous comparisons of multiple models for the purposes of model assessment (how good is the model compared to others) and model improvement (how progressively good vs better vs best is the model compared to 50/50 chance, when retaining or discarding various variables in the model while trimming the model to contain ONLY significantly predictive variables, toward the end point that all variables in the model are independently and significantly predictive with an overall model probability AUROC Curve of greater than 75% (AUROC Curve = 0.75 or higher). A good model. A strong, excellent model would have an AUROC Curve of 95% (0.95) with 95% of the model probability being explained by a set of independently significant factor variables in the model. A perfect model would be 100%
If your Dependent Variable(s) are a distribution of a scale variable value, it will necessarily need to be dichotomized into a binary (yes/no) variable for use in a Binary Logistic Regression model, but CHAID will handle distributions of raw scale dependent variables, and may actually provide insight as to ideal an "cut point" location for dichotomization in a continuous scale variable distribution. Run CHAID first, pay attention to where the branches are formed in the continuous scale dependent variable (CHAID uses the interaction of the independent variables with the dependent variable to form "branches" that fork at "nodes". The branches are formed around an ideal "cut point" in the Dependent Variable, and this can serve as an initial operational definition of the Binary Dependent Variable that is required for all BLR models. It is a recursive process, but it quickly leads to trimmed BLR models from a large laundry list of independent predictors with unknown and invisible higher order interactions.
Financial data can provide rich continuous scale level of measure, and categorical and ordinal data (type, class, commodity, frequency, rank) can provide actionable insight regarding preference, order of purchase, location of purchase, and other very interesting and informative patterns of transaction that have domain (clinical, financial, legal, business) significance.
CHAID, SPC Charts, and BLR would be the models and methods that I would use on first pass at analysis of financial transaction data that you describe. I would use a Week-Year time variable for the horizontal time axis of the SPC Weekly time-series charts. I would use Month-Year for Yes/No dummy coded variables to test and control for any effects of seasonal variation over a minimum two-year time period in the scope of data time-frame. Main threat to validity of data involving human activity is seasonal variation itself, and the only way to control for it is too collect data over a minimum two-year time frame. Other wise, you will need two parallel groups at the same time, one control, the other experimental if you do not have the luxury of a full two-year time frame. Even then effects of seasonal variation, especially in a time-series, is a formidable challenge and can be a "show stopper" (as COVID-19 seasonal effects) have turned out to be for any and all human subjects research data collected within the 2020-2022 and beyond time frame.
Hope this helps.
Richard E. Gilder, MS RN-C
Bioinformatics Scientist
The Gilder Company
TTUHSC Adjunct Faculty
  • asked a question related to Information Retrieval
Question
1 answer
If you have for example, 5 different samples, that were sequenced, using 3 different regions, but have realised the one gene region is not powerful enough to resolve it to species level, so you add two more genes to better the topology of the phylogenetic tree
Relevant answer
Answer
Gosiame Boitumelo Malepe A phylogenetic tree is constructed in four separate steps: (Step 1) find and get a collection of homologous DNA or protein sequences, (Step 2) align those sequences, (Step 3) estimate a tree from the aligned sequences, and (Step 4) show that tree in such a way that the important information is clearly conveyed to others.
  • asked a question related to Information Retrieval
Question
5 answers
Would you please share your experience from where I can collect firm-level green patent citations for green innovation? Please mention the name of the source ( I prefer free one).
Relevant answer
Answer
Thank you so much, Dr.Tim John Hewson for sharing your experiences. Dr. Is Espacenet has a firm-level patent citation?
  • asked a question related to Information Retrieval
Question
3 answers
Hello everyone. I am doing my dissertation on factors affecting SME internationalization in Bangladesh a comperative insight between UK and Bangladesh. and I decided to complete my paper using secondary data sources. but I am stuck in deciding which data analysis I should use. i am using interpretism philosophy and an inductive approach. and my strategy is the case study.
Relevant answer
Answer
Your question says you need to take courses of research methods atleast for one year. You are not yet ready to do research even under supervision.
  • asked a question related to Information Retrieval
Question
2 answers
Is it possible to conduct a moderation analysis when all variables (X, Y and W) are categorical (dichotomous)? I believe I have identified a moderation effect by splitting the data file based on my moderator (W) and observing different effects of the predictor (X) on the outcome (Y) but when I explore the interaction in a PROCESS moderation model (Model 1), the interaction term is not significant. Am I missing something? All variables have been dummy coded.
Relevant answer
Answer
I recommend that you use the free statistical program JAMOVI which allows the inclusion of multiple moderators and mediators.
  • asked a question related to Information Retrieval
Question
2 answers
Is there any free online software to process the fluxomics raw data files.???
1. To calculate 12C and 13C ratio
2. Pathway mapping
Relevant answer
Answer
OptFlux3 is open source and "might" be modified to meet your needs.
  • asked a question related to Information Retrieval
Question
2 answers
Educational purpose
Relevant answer
Answer
Rizwan Farook Hard copy maps, aerial images, remotely sensed imagery, point data, survey samples, and existing digital data files are the most frequent generic sources of spatial data.
  • asked a question related to Information Retrieval
Question
28 answers
Dear All,
I am looking for some Partial discharge (PD) datasets to download. I would appreciate if you can mention some data sources from where I can download the PD datasets.
Regards,
Anis
Relevant answer
Please share with me: soronilameitei@gmail.com
  • asked a question related to Information Retrieval
Question
8 answers
I am looking for Nitrogen, Potassium, Phosphorus, Organic carbon stock , pH value, Added Nutrients features in soil dataset.
Kindly suggest me the relevant data source.
Relevant answer
Answer
@ Manoj, you may get it from Soil & Land Use Survey of India (Govt. of India). You can also download from FAO Soils Portal, GloSIS Global (Beta), European digital archive on soil maps (EuDASM), International Council for Science (ICSU) World Data System and GEOSS (Global Earth Observation System of Systems) portal.
  • asked a question related to Information Retrieval
Question
1 answer
I want to do MD simulation with LAMMPS on calcite, magnesite, silica and other mineral to figure out their minerals. But, firstly I need to create data file that contains coordinates, bond, angles, charge and other coefficients. I tried with atomsk to convert cif files into lammps data files, but, there was not found any parameters for bond, angle and charge. I need help to solve this problem. Would you like to suggest any software or direction by which I can make the data files with those coefficients? Thank you.
Relevant answer
Answer
Have you checked the recommendations from https://www.lammps.org/prepost.html#builders ? In addition to atomsk, they list 9 different molecular builders. At least Moltemplate mentions bonds, angles and dihedrals explicitly, but I would assume most of them can produce all necessary input for LAMMPS, being listed on the official site.
  • asked a question related to Information Retrieval
Question
9 answers
Hi,
I am simulating a centrifugal compressor using the single passage method in Ansys.
The case file works correctly in Ansys 19, however, when I open the converged case and data file with Ansys 2021 R1 and continue the iteration process, a sudden jump occurs in the residuals and the final results also changes and wrong.
Can anyone help in this regard?
Thanks in advance.
Relevant answer
Dear Morteza,
Ansys solver is going to be slightly different in each new version in terms of its input parameters. Once you heat the solve button, ANSYS gathers all these parameters from your model and creates a file (.dat) and consequently sends it to its solver.
Therefore, if you just resume an analysis which was started in an "old" version in a "new" version, then the new solver expects parameters from the old package that are slightly different or even doesn't exist. This causes a significant interruption in the solver.
I suggest you start over your analysis in the new version to avoid having inconsistency in solver parameters which can lead to a significant error in your simulation results.
  • asked a question related to Information Retrieval
Question
2 answers
I urgently need the reference XRD data file for beta molybdenum carbide. I also need a free source to get diffraction data other than COD.
Relevant answer
Answer
Dear Bastien Leclercq,
Thank you for your help.
  • asked a question related to Information Retrieval
Question
2 answers
I am wondering is there any available EUV brightpoint database from the SOHO and SDO period? This catalogue can be a data file, fits or jpg.... I want to make EUV brightpoint synoptic maps. I am also interested in plages, filaments, faculae and/or prominence databases.
If someone knows these types of databases then would you be so kind to post related links?
Thank you in advance.
Relevant answer
Answer
  • asked a question related to Information Retrieval
Question
5 answers
I want to know how to run an MD simulation on a metal or substance for getting to know about its stress, strength energy, deformation, etc. for doing research on it.
Relevant answer
Answer
This might help you with the initial simulation.
It would help if you also studied published articles based on your research focus. For example, you can see the below article for stress analysis using LAMMPS.
  • asked a question related to Information Retrieval
Question
3 answers
I obtained an EIS data when I am doing research on biopolymers for corrosion inhibition. But I can not find a good fit of equivalent circuit. Could any body help me with it?
I have attached the data file in which the first column is the frequency, the second column is the Z' and the third column is the Z''.
Relevant answer
Could you provide more details about your surface morphology ? ... Is it only a metal covered with this biopolymer ? ... The choice of an equivalent circuit depepnds strongly on what is happening on your surface (electrochemical and physical phenomena) and the state of your surface (porosity, coating, protected and unprotected areas ... ). Your data is corrected for the ohmic drop or not ?
  • asked a question related to Information Retrieval
Question
3 answers
Hi. Everyone. I have a problem with WordSmith tools on doing the data collection of 4-word Lexical Bundles that I only see the number over there but cannot check the data source like I use AntConc to do 4-word bundles that can be shown the LBs types. So how can I check the data source of 4-letter words? For example, Text 1 has 135 LBs was calculated and where can I see the whole LBs types? I spent a whole day finding out, but nothing can help at all. I am so confused about this tool now. If anyone can help me, I will appreciate it.
Relevant answer
Answer
Laoxin Qin To begin creating a word list, push the WordList button on the main Controller. When WordList loads, choose your texts, and you'll see something like this. Here, we'll create a single simple wordlist based on eight text files from the play Romeo and Juliet, so click Make a word list immediately.
  • asked a question related to Information Retrieval
Question
5 answers
Hello all,
Does anyone know how to generate LAMMPS Data files from material studio Coarse grained beads?
I generated a CG model using material studio's Mesocite Module. But, I am not able to generate the LAMMPS data file from it. When I am exporting the file, the file does not contain any info.
Any suggestion would be highly appreciated.
Thanks.
Relevant answer
Answer
CG Artists (also known as computer graphics artists) create 2D and 3D art, usually for Cinema, Advertising or animation movies. A CG artist's work usually revolves around finding balance between artistic sensibilities and technical limitations while working within a development team.
  • asked a question related to Information Retrieval
Question
7 answers
I have a data set that contains a text field for approximately more than 3000 records, all of which contain notes from the doctor. I need to extract specific information from all of them, for example, the doctor's final decision and the classification of the patient, so what is the most appropriate way to analyze these texts? should I use information retrieval or information extraction, or the Q and A system will be fine
Relevant answer
Answer
DEAR Matiam Essa
This text mining technique focuses on identifying the extraction of entities, attributes, and their relationships from semi-structured or unstructured texts. Whatever information is extracted is then stored in a database for future access and retrieval.The famous technique are:
Information Extraction (IE)
Information Retrieval (IR)
Natural Language Processing
Clustering
Categorization
Visualization
With the increasing amount of text data, effective techniques need to be employed to examine the data and to extract relevant information from it. We have understood that various text mining techniques are used to decipher the interesting information efficiently from multiple sources of textual data and continually used to improve text mining process.
GOOD LUCK
  • asked a question related to Information Retrieval
Question
5 answers
From same data source/work, is it okay to publish a correspondence/letter to the editor first using preliminary findings then after completion of whole work publishing the original article using more indepth analysis of data in different angles of presentation and detailed discussion?
Relevant answer
Answer
Dear Md. Rabiul Islam, I think you have got your answer very clearly from the clear explanations of Cecilia-Irene Loeza-Mejía and Rahul Sharma. I also think that if your research objectives and methods are different, you can use the same data source/work. And the results of the next article must be different from the previous one.
  • asked a question related to Information Retrieval
Question
3 answers
We are conducting a study about COVID-19 Burden in Romania in 2020-2021, and we need the number of cases treated in hospitals in Romania. Can you indicate an open-source data in electronic format about this parameter?
  • asked a question related to Information Retrieval
  • asked a question related to Information Retrieval
Question
3 answers
I have row data files; need to find amplitude and phase of the signals. Is it possible to find it using the origin.
  • asked a question related to Information Retrieval
Question
2 answers
Hi,
I need epigenetic data file run through the hovarth clock. DNA Methylation Age Calulator, but did not find any description how it should be done?
Please can you help me to find link to describe the process of running the data?
Relevant answer
Answer
The website can be found here: https://dnamage.genetics.ucla.edu
Look at the 'TUTORIALonlineCalculator.pdf' file on the page to understand how to submit your data as a CSV file and how to interpret the results.
Alternatively, the R code is available. I think it's linked to in the supplementary data of Horvath's original paper, if I remember correctly.
  • asked a question related to Information Retrieval
Question
5 answers
I developed an approach for extracting aspects from reviews for different domains, now I have the aspects. I want some suggestion on how to use these aspects in different applications or tasks such as aspect based recommender system.
Note: Aspect usually refers to a concept that represents a topic of an item in a specific domain, such as price, taste, service, and cleanliness which are relevant aspects for the restaurant domain.
  • asked a question related to Information Retrieval
Question
3 answers
I am trying to use Zview software but when I try to open a data file it shows me this error message. what should I do to solve this problem? it's urgent please.
Relevant answer
Answer
The error indicates that the software is in trial mode, that is, it needs a license to be able to use it. You could take a look at ZView alternatives:
Best Regards,
Cecilia-Irene Loeza-Mejía
  • asked a question related to Information Retrieval
Question
6 answers
Kindly suggest how do I get the data file of Ion Ranges in SRIM Program. I want to plot the data in another plot program for analysis.
Relevant answer
Answer
  • asked a question related to Information Retrieval
Question
9 answers
I'm doing data analysis for data that has two sources of diversity with replicated data (source of data is distributed in group). Unfortunately the data are not normally distributed. The alternative test that I got was the Friedman test but I found that the Friedman was only for non-replicated data. Are there any other non parameter test i can use?
Relevant answer
Answer
What assumptions failed that decided you go with non-parametric methods? Just curious to know because parametric ANOVA is pretty robust to failing assumptions.
  • asked a question related to Information Retrieval
Question
2 answers
Hello everyone,
As many of you are aware, Industry 4.0 and smart manufacturing is widely gaining traction among many manufacturing practitioners.
One of the aims of smart manufacturing is to bring in more real-time intelligence to shop-floor and manufacturing planners, by employing different shop-floor data sources (For eg, smart products, advanced equipment, and machine sensors, collated through the Industrial Internet of things, RFID gateways, etc). The data collected from these shop-floor data sources can be processed by means of Artificial intelligence-based algorithms ( machine learning, evolutionary algorithms, etc) to produce optimized production schedules, process plans, service schedules, and maintenance plans. The application of such techniques has been researched by several researchers and numerous publications are already available.
However, at the same time, central to the functioning of any manufacturing industry are Enterprise resources planning packages, encompassing all the functions of a manufacturing business, ranging from procurement to production planning and control to service management and even auxiliary support functions such as accounting and packages.
As a result, there are two different software systems that can benefit manufacturing firms:
1. AI-based smart manufacturing tools which seem promising in improving production efficiency: Such packages are more manufacturing operation centric
2. Tried and tested ERP work packages, developed by numerous software firms: Such packages can focus on both the business planning and manufacturing operations management.
The questions that I often wonder about in this regard are :
  • Have AI- algorithm-based smart manufacturing tools been integrated into existing ERP software?
  • What is the level of maturity of ERP work packages with respect to AI-based intelligence algorithms ( particularly with respect to integration and hosting of AI-based work packages, API access to databases, software architecture)?
Many ERP work packages support multiple scheduling rules for a host of production scenarios such as flow manufacturing, make to stock, make to order, job shop- etc.
  • But do they actually employ AI Solvers, such as genetic algorithms, search-based algorithms, Neural network models, reinforcement learning-based algorithms?
From my initial assessment of the market, I understand that we are staring at a situation where the development in these two work packages progressing is in two nonintersecting planes, with two separate software packages being the only way out- one for AI-enabled manufacturing execution and the other for ERP systems?
In that case, I'd be keen to know the market feedback-
  • Are industrial engineers of today willing to accept the need for two separate software packages for manufacturing scheduling and control, considering that liaising with software systems is not their core job ?
I've found a few articles on Intelligent ERP systems, but they are pretty generic, and mainly focus on the need for ERP systems to integrate cloud and mobile-based support and automated inspections, but do not discuss much on AI solvers.
A search on google scholar does not reveal much either, with many articles proposing architecture for ERP of the future, integration with smart agents, etc, without discussing a lot on existing maturity levels and capabilities.
Looking forward to your valuable answers!
Relevant answer
Answer
ERP systems, e.g. SAP, Sage, Oracle, IFS, and Dynamics, are well developed and adopted by enterprise companies. I trust Big data analytics and related technologies have been utilised for advanced planning. However, the cost and time for ERP implementation and integration is still a severe challenge to many companies. So, Google scholar may not show ERP capabilities and innovations that exist!
  • asked a question related to Information Retrieval
Question
4 answers
The main problem I always encounter with external packages or internal commands is that bonds that pass through the boundary of the simulation cell, or atoms that exist exactly at the wall of the cell(super common for inorganic structures, polymers, MOFS, coordination polymers, and all the interesting stuff...) always get problematic in some way. Bonds that pass through the wall usually break), other times atoms that exist on the wall get duplicated while they really shouldn't.
I have used topotools, Lammps-interface, Atomsk, OVITO and other external software with topotools and Lammps-interface seeming the most promising from a topological point of view, but again I get the same problems with them
Relevant answer
Answer
It seems I kind of found a solution to my problem. A combination of ASE, Lammps-interface, OVITO, and some manual modifications, that gives very good results.
Methodology:
1)Obtain cif file
2)Parse cif file to P1 structure using ase with the command: ase convert input.cif output.cif
3)(optionally but for me is a must) Because lammps interface oversimplifies topology/force-field parametarisation one can modify manually the output.cif and replace some atoms of the same type with dummy ones (of other elements) in order to make lammps-interface assign different topological parameters (bond and angle types) that you will manually replace in a later manual modification of the lmp.in or lmp.data files with the correct ones. (For example I replaced some of the metal coordinated oxygens with sulfur and selenium atoms and replaced them in a later stage)
4)generate lmp.in and lmp.data files using lammps-interface for 1 unit cell (beware of getting the cutoff of non-bonding interactions smaller than default something like 2-3 Angstroms will be ok and change it later manually).
5)Add partial charges to the atoms (I do it by Periodic DFT using Quantum Espresso for 1 unit cell, extract valence charge density, and using the chargemol code, I calculate the DDEC6 partial charges for each atom in the unit cell and transfer these charges manually to the lmp.data file)
6)Modify all necessary topology parameters (for example: Replace bond parameters of M-Dummy with M-O, replace Lennard Jones interactions of dummy atoms with the normal ones, do the same for angles and dihedrals if needed). Once you have refined the unit cell simulation parameters and topology as much as you want/need->
7)Add data.lmp file to OVITO
8)Replicate cell using OVITO (OVITO does a very good job on making replications and keeps the topology intact. It will cause no problems in the boundary after duplication)
9)Export the new data using OVITO as a new lammps data file for example lmp(supercell).data
This will duplicate everything properly, keep the topology intact, transfer partial charges, bonds angles, and dihedral topologies in the right way.
10)OVITO forgets to transfer the force field parameters (all of them) so open the unit cell lmp.data file, and just copy and paste the force field parameters to the new lmp(supercell).data and here you go!!!
This procedure worked for me with a new 2D Proton conducting Metal-Phosphonate coordination polymer that I synthesized and am trying to simulate. Right now I have a perfectly working classical NVT MD simulation of 2x2x2 supercell of my compound. Gonna try to simulate some proton conduction now.
  • asked a question related to Information Retrieval
Question
1 answer
I am new to molecular dynamics. I am currently using the Material Studio to prepare the data file and then convert the car.file into a LAMMPS data file.
But when I use the amorphous cell, some molecules are actually across the boundary of the cell.
I want to confine the molecules inside the cell boundary.
Or I want to know how to make the cell bigger.
I know I can add a vacuum or revise the length of the cell but that only add space for one side.
Appreciation for any response
Relevant answer
Answer
All the molecules are actually inside the simulation cell. A few of them seemed to be out of the box, but due to periodic boundary conditions (PBC), they are actually present inside the simulation cell. If you want to see it, then change the display style lattice to the In-cell option.
  • asked a question related to Information Retrieval
Question
3 answers
Hi
In past I was using MASCOT to search, identify and explore the proteomics data. I am looking for any windows based free software to analyze my big proteomics data files mostly in mgf format.
Relevant answer
Answer
Thanks everyone. I have downloaded the Maxquant...let's see how it goes....
I am looking for both quantification as well as identification...moreover looking for PTMs too
  • asked a question related to Information Retrieval
Question
9 answers
I need the cif diffraction data file for Nb2O5. The specification for this compound are given below:
Chemical Name: Nb2O5
Crystal Structure: psudo-hexagonal
a:3.607 b :3.607 c:3.925
I cannot find the cif file from Findit or ICSD. I would be really grateful if anyone share it.
Thanks a lot.
Relevant answer
Answer
Gustavo Henrique de Magalhães Gomes I've seen the structure file you've shared for us. Do you mean that the cell of TT-Nb2O5 isn't actually hexagonal but orthorhombic?
  • asked a question related to Information Retrieval
Question
6 answers
My name is Huiyeon and I am using STRUCTURE 2.3.3 software for or Population structure analysis. But I have a problem.
I have genotype data scored as 0 and 1 for SSR marker in formats available on the NTSYSpc program.
First, I converted my Excel files into TxT file format (Tab delimited).
Next, I tried to import the file into a new project.
And I specified according to instruction and number of my sample is 105.
When I finish setting and want to proceed, error "Bad Format in Data Source: Expect 27 data entries at line 1" occurs, and clicking "OK", a window says "The data file is expected to have following format: 1 row with 24 entries(marker name)/1 rows with 27 entries (data)".
Though I turned back to the setting and modified Number of loci several times, almost the same error occurs.
How can I convert my genotype data file into STRUCTURE input data file for Population structure analysis?
Kindly advise.
Thank you.
Relevant answer
Answer
It seems an error in your data format. You have to replicate your data according to the ploidy level of the species which on you are working. If you want, then I can help you to format the data.
Regards
Chet Ram
  • asked a question related to Information Retrieval
Question
2 answers
What is the most reliable accurate data resources for solar flux indices?
Relevant answer
Answer
Dear Dr
The below websites might be interesting for you in order to reliable solar flux data.
  • asked a question related to Information Retrieval
Question
4 answers
Hai.
I am trying to use Infocrop model for climate change effect. I have imported csv file of daily weather data from 1971-2002 into infocrop. It has successfully created the 32 files and CLI file also. When I create the project and use the weather file, it does not list out the years in the pull down. So please tell me the what is the reason and how to solve, or is their any other way to import the weather file,
Thanks in advance.
Meti
Relevant answer
Answer
Facing the same problem
  • asked a question related to Information Retrieval
Question
3 answers
There are few steps to make heatmap of your qRT-PCR data (fold change or relative quantification) using R.
Data file preparation:
Make excel file of your data in which your will place your gene of interest in column and your treatment or conditions in row.
Save the file in *csv extension.
Import data file in R:
By using following codes, import your data file into R,
data2 <- read.csv("data1.csv")
~ data1.csv will be file name your data file your created in excel and data2 is the name of your data in R. You can use your own names instead of data1 or data2 and you can even give your data a single name at both places.
When you will import the data, you will see first column composed of serial numbers. We need to replace the numbers with the names of actual column of your data that contain your gene of interest. To do this use this code:
rownames(data2) <- data2$Name
~ Name is first column
This will replace the serial numbers with your first column. But now you have two columns with your genes of interest. To remove duplicate, use this code:
data2$Name <- NULL
Now your data is ready to create heatmap.
Developing heatmap:
First create matrix of your data by using following code:
data2 <- as.matrix(data2)
Now install a package to create heatmap "pheatmap" by following code:
install.packages("pheatmap")
after installing you will call that package every time when you want to use it by following code:
library("pheatmap")
Then give a command to make heatmap of your data by following codes:
pheatmap(data2)
Usually we show fold change/relative quantification value inside our heatmap to add them modify your code in the following way:
pheatmap(data2, display_numbers = TRUE)
- You can customize your heatmap in many ways. Contact me any time if your any help.
Relevant answer
Answer
Good luck