Science topic
File Systems - Science topic
Explore the latest questions and answers in File Systems, and find File Systems experts.
Questions related to File Systems
I measured some samples for Raman spectra (Renishaw). The obtained data are in .wdf and .wxc format. These data can be accessed by WIRE software (Renishaw), which is not open-source/ free. Is there any other way I could read the data?
How edge computing can be useful in structured DBMS file systems? can anyone have a paper or article regarding this?
System information
- OS Platform and Distribution (e.g., Linux Ubuntu 18.04): Ubuntu 20.04
- Python version: 3.6
- Installed using virtualenv
- CUDA/cuDNN version: 11.5 / 8.1.0.77
- GPU model and memory: RTX 3090 24GB nvidia driver 460.39
- TensorFlow version: 2.4.0 pip install tensorflow-gpu==2.4.0
Describe the problem
Installed cuda 11.2 and cudnn 8.1.0.77. Faced the following problem when I run train.py
Could not load dynamic library 'libcusolver.so.10'; dlerror: libcusolver.so.10: cannot open shared object file
Solved the problem
- place libcusolver.so.11 in the tensorflow site-packages of the virtual environment created worked properly
- followed the link to solve the problem
- Manage a link between libcusolver.so.11 and libcusolver.so.10 using a comand 'sudo ln -s'
$ ln -s /usr/local/cuda/lib64/libcusolver.so.11 ~.virtualenvs/(env name)/lib/python3.6/site-packages/tensorflow/python/libcusolver.so.10
This fixed the problem for me
Some 3D lossy compression software can achieve more than 70% compression, whilst lossless compression achieves less than 20% and needs to be decoded to be able to use it, what will it mean to you/company(blender, unity, animation companies, etc...) to have a lossless compression software that will achieve more than 50% on all types of 3D files(STL, OBJ, AMF, 3MF, FBX, PLY, etc...) that compress the files using succinct data structure thereby no need for decoding of the file?
Definitely by compressing you'll save lots of space and transmitting the file online will consume less bandwidth, hosting the files will be cheaper, etc... well as some might have guessed, I have developed the software.
1. So I just want to know the economic impact of such software would it run in $Millions or Billions?
2. What companies would be interested?
Hi, everyone,
I just calculated a pure water-box(32 molecules pre-equilibrated by LAMMPS) to learn how to simulate a NVT ensemble by VASP, but unfortunately, I cannot get a converged energy profile(shown as the figure). It keeps increasing! Could anyone provide some suggestions?
Besides, I noticed that it is the potential energy of the Nosé thermostat keeps increasing, while the F or E0 converges well.
Here is my INCAR file:
SYSTEM = Test
LSCALAPACK = .FALSE.
#Start parameters
NPAR = 6
PREC = Normal
LREAL = Auto
ISTART = 0
ICHARG = 2
#Electronic relaxation
ENCUT = 600
ALGO = Fast
NELM = 300
EDIFF = 1E-5
NELMIN = 5
#MD parameters
ISYM = 0
IBRION = 0
POTIM = 0.5
NSW = 30000
TEBEG = 300
IWAVPR = 11
#NVT canonical model
ISIF = 2
MDALGO = 2
SMASS = 0
#DOS related
ISMEAR = 0
SIGMA = 0.05
#Switches
LWAVE = .FALSE.
LCHARG = .FALSE.
IVDW = 11
Thanks a lot.
With those three different OS at hand, I wanna know how they would differ in terms of 'Process Scheduling', 'Inter-process communication', 'Memory Management', and 'File Systems'. Is it as simple as, process scheduling for Linux is done with Completely Fair Scheduler (CFS), while Windows uses ?? and Solaris with ?...…or is there more to it. Someone can help on how I would do it for the others (IPC, etc.) ?
How can we use IPFS with BlockChains to solve different problems.
Consider a configuration where the name nodes have plenty of memory and storage matching the data nodes in a Hadoop cluster. There are 2 Name nodes. Can the name nodes also be configured as Data nodes?
i assume if the Name nodes are virtualized, the answer is yes. Either way, I like to know what others have done in this type of situation.
What id diff. b/w Distributed file systems and distributed database?
I want to log user activity in a way to determine how they access their files on daily basis. For example, student have book in pdf and he read this file and other files during learning process. Log data will be used to model of users. Any ideas or suggestions are welcome, it would be best if there is some built in feature or tools inside Operating system so user stay unaware of tracking their behavior.
Is there any difference between them?
I'm calculating the lowest-energy triplet state (ISPIN=2, NUPDOWN=2) of a system comprising an organic molecule attached to a silicon surface. After performing an electronic minimization, I'm seeing the following output in my OUTCAR file for this system:
eigenvalue-minimisations : 38300
total energy-change (2. order) :-0.9235746E-04 (-0.2775273E-06)
number of electron 1086.0000388 magnetization 1.9999998
augmentation part -42.1581394 magnetization 0.1063087
The "augmentation part" is registering at roughly -42, which doesn't make a lot of sense to me. Only the last iteration of the minimization is shown above, but this negative value for the "augmentation part" occurs during all self-consistent steps after the initial NELMDL=-5 non-consistent steps. In my previous work, I've never seen this number be negative. I've also only done non-spin-polarized singlet states in the past, so perhaps the above is caused by the spin-polarized triplet nature of the calculation? I'm not sure. Any information on why this occurs and whether it's OK to see this would be greatly appreciated.
I have tens of thousands of individual scans in proprietary file formats, and I want to make these public. I need a format that is free and open, or to make my own.
Our proprietary software offers a CSV option, but doesn't export all useful data to the file. In addition, the CSV file it creates is more like two spreadsheets, with the second half having the per-channel photon counts.
I've considered using XML because it is both machine and human readable. My only concern is that XML is bloated. XML has the added benefit of being readable over a web browser, and can be quickly converted to almost any language, including JSON.
Microsoft INI format is also machine and human readable, but INI is fairly phased out. Software writers still have full access to INI functions though, so I wonder if this is still a viable format. INI also converts well to object notation.
Both INI and XML could better represent two spreadsheets worth of different-typed content in a single file than a CSV.
What are your thoughts?
I have tried reorienting and fixing the following MnI(6,6'-morpholino-2,2'-bipyridine)(CO)3(acetonitrile) complex such that it would optimize in gaussian and no matter what I do to it I always get the same error:
Error in internal coordinate system.
Error termination via Lnk1e in /share/apps/gaussian/g09/l103.exe at Thu Dec 20 16:57:56 2018.
Can someone suggest a method to fixing up my input file such that this error will not occur?
I will include the input and output files for this system with my question so feel free to browse them and see if there are any major corrections that I can make.
I have a modeled structure of RNA with 25 nts. I have renamed the atoms as per the rtp file of the amber99sb-ildn force field. All the steps upto energy minimisation ran without any error. The equillibration step with NVT of 200ps is running extremely slow. I have already run a protein (90 aa long) simulation with exact same parameters which completed NVT in 10 mins. I have no clue where things are going wrong.
p.s. I found it odd that the topol.top file generating after the command: gmx pdb2gmx -f rna.pdb -o rna.gro (with -ff amber99sb-ildn and -water tip4p) is showing system name as 'protein'!
The last few lines of the topol.top file is:
[ system ]
; Name
Protein
[ molecules ]
; Compound #mols
RNA_chain_A 1
Due to this rna_solv_ions.gro shows system name as 'Protein in water'
I don't know if it is relevant to the problem but I found this odd. As if, pdb2gmx cannot recognize my pdb file as a proper RNA molecule!
Kindly help me to understand the issue and resolve it.
Thanks in advance.
Java is intended to be used in networked/distributed environments. Toward that end, a lot of emphasis has been placed on security. Java enables the construction of virus-free, tamper-free systems. ...Source: Core Java 10th edition....
The main problem is the fact that malicious code can cause its damage because it has gained unauthorized access to system resources. For example, a virus program might gather private information, such as credit card numbers, bank account balances, and passwords, by searching the contents of your computer’s local file system. In order for Java to enable programs to be safely downloaded and executed on the client computer, it was necessary to prevent them from launching such an attack. Java achieved this protection by enabling you to confine an application to the Java execution environment and prevent it from accessing other parts of the computer. ...Source: Java Complete Reference 9th edition....
Python has become popular for business, scientific, and academic applications and is very suitable for the beginning programmer. There are many reasons for the success of Python. Python has a much simpler and cleaner syntax than other popular languages such as Java, C, and C++, which makes it easier to learn. The Python programs in an interactive environment, which encourages
experimentation and rapid turnaround. Python is also very portable between computer systems. The same Python program will run, without change, on Windows, UNIX, Linux, and Macintosh. ....Source: Phyton for Everyone 2nd edition....
I have recently gained interest in big data/data mining, and have been reading up on my own. When I surfed through Internet about it, I found mechanisms like map-reduce jobs,distributed file system etc.
I want to know what are the major algorithms that are used in this field for real time problems.
Does/can HDFS manage files across multiple geographically dispersed datacenters? What distributed file systems can be used for this purpose?
I am trying to perform a simulation of polymer brushes based on PNIPAM with GROMACS code.
Generating a coordinate file for the system of interest is not problematic, but creating a topology of this non-linear case is complicated (default gromacs tools are dedicated to linear systems like peptide).
Could you recommend the best tool for this problem (topology generation)?
I have ChIP-Seq data I have been analysing with Homer. I wanted to do Irreproducible Discovery Rate (IDR) analysis on my replicates. So I downloaded the homer-idr and IDR program onto a MAC OS X. I have installed both but IDR will only work when I remove homer-idr. When i have both installed and I launch python from terminal:
>>help('idr')
it gives me the info for homer-idr not IDR. When I remove homer-idr it gives me the correct info for IDR. Both have the same package name if that is my issue.
Does anyone know how to fix this?
Cheers, Ben
I am actually trying to extract data from SEC Edgar filings, however I see that building parsers for each form is quite exhausting, in addition not all those filings are of same format even though they are from the same form say for eg. 10k filing.
I am intrigued if some one can direct me in the right direction.
I want to do some waveform analysis in SAC. I am working in rtqt view with displacement records of three component data in ASCII format. Now I want to perform some phase picking and particle motion plotting using SAC. How can I change my data files in SAC format? Any suggestion regarding this conversion would be helpful.
I have downloaded satellite data from TOMS and OMI Sites. It is in .he5 and hdr file format. How can I open such files?
I require a ready to use VM of Lustre File system, can any one help me for the same? Thanks in advance.
Compression is used for the audio and video file from olden days. In a database how it can be handled in efficient way?
Actually, suggestions are needed as to whether it will be applicable for HDFS or any other DFS.
I have been trying to extract meta data from various types of files. I am looking for attributes that are available for any type of file format so that I can use them for clustering. I need suggestions for selecting attributes. Please suggest me some meta data that is available for all kinds of files. If the data are numeric that would be better. Thanks.
Hi Dear friends,
Is there any difference between cloud file system and cloud storage system?
and what are strategies used in dynamic replication system on cloud storage
I had X.dat file of 1.5GB size while converting to mat file using the statement
a=dlmread('X.dat',' ');
I got an error as Error using dlmread (line 139)
Mismatch between file and format string.
Trouble reading number from file (row 1u, field
1u) ==>
Is there any other method or there is any delimiter I can use?
Dear colleagues,
running computational software in parallel on 24 cores has an reasonable data reading time, but by increasing the number of cores up to 80 cores it takes an excruciating time to read data.
my opinion is that if I use parallel file system, I may overcome this matter.
What does PFS do? Does it help me? how much?
Is there any user-friendly, windows OS PFS,and prefered to be open source,and have a good installation guide?
To be honest I couldn't install orangeFS, because of installing servers which needs linux
I recall a paper some time ago that purported to estimate the cost of deleting a file (e.g., a picture). The argument was that it was cheaper to add storage than to delete files. I am not able to track this down, but my interest in the question remains. I am aware of studies that estimate the cost of erasure from the point of view of physics, but that isn't what I was looking for. Perhaps as a alternate some sense of consumer behavior with regard to adding storage vs. deleting files would be helpful.
I am reading a file system related paper. It says a concept "directory indirection". Can anyone explain to me how it works? And also, what is the difference and what are the benefits compared with the classical file system structure?
I have a HEALPix file and want to convert into normal fits file using a Python. Can anyone show me how this is done in Python?
What are the most fundamental and important papers and books about various kinds of features and their extraction for retrieval and classification of multimodal files containing text, images, audio and video?
The mapping between semantic data (e.g. file systems and namespaces, databases, object stores) and the devices that store the associated bytes and blocks that compose them has been one-way by design - from semantic data through pointers (e.g. inodes) to blocks of bytes, but the reverse association is rarely used, except for debug. In semantic storage, the storage controllers (and in some cases devices themselves) would know that a block belongs to a specific object, file, table/record/field and this reverse association could be used by new applications for performance optimization, security monitors, and data protection, for example. The idea is not to revolutionize or change storage, but rather allow for reverse mapping so lower-level features can be semantically aware and so new applications like intrusion detection systems can know that block access is suspect, sub-optimal, or requires cache updates.
I find problems in open a 3d image file by Matlab to find a solution for the restoration of the broken the ceramic fragments data that scanned in 3-D scanner. What is the most suitable method for open and read 3d image file by Matlab.
With Best Regards.
I have a .txt file that contains information in the following pattern :
The data is separated in the form of
255,205,0 102,235,39 206,89,165 ....... (that is, 3 uint8 integers separated by commas, and the groups of 3 separated by whitespaces). There are a total of 30*60 = 1800 triplets of numbers separated by commas and the triplets are separated by spaces.
Basically, these are pixel intensities of the 3 channels in an RGB image. I need to store them in a 2 dimensional array such that the first element of each triplet goes into the 1st column, the second element into the 2nd column and the 3rd element into the 3rd column. Effectively, at the end of this operation, I should have a 2 dimensional matrix of 1800x3 size.
Please help me with this. I have attached a sample text file. Thanks in advance.
How can we design a file system for storing the data? Will it be fine if we make a dummy program in C/C++ with a predefined size of window that accepts input of the file of 5KB and stores in it by maintaining the address table of each data block as well as the hierarchy. I know that in real, file system is a complex thing and more than this dummy program.
If there is any good material or simulation tool available for this then inform the same.