My problem is my edit distance values are stored in a 2d array of ints and exported to a .csv and the ones at the end are rather LARGE.
The the edit distance max should be around 2000 but i am getting values of 1-100million, the weird thing is that I have checked back through my function and tested various parts of it and still dont understand where it is going wrong i thought it could be my memcpy and memmove parts but i have had no luck.
I built a C++ static library using WatCom IDE. It compiles without error, but running WLIB on it produces the msg:
Warning! Library contains no external symbols My WatCom compiler host is windows, & target is DOS.
If I convert the Library to an .exe program, by adding a main(), it executes fine, So there seems to be something I'm missing in the IDE settings for building a .lib.
I have a program where I want to export the .lib file in a "Lib" folder, and the .dll file in the "bin" folder. How can I choose where the lib file and .dll file go?
I'm building a cross-platform library which links to some other 3rd party libraries at run time (i.e. on Windows, the other libs will be available as DLLs whereas on Linux / OS-X etc they'd be shared objects, which are similar). For the sake of argument, one of those libraries is called "jack".
Obviously, our app can't guarantee which version of the other libs will be on the user's system (or even that they'll be installed at all). So our code is littered with statements like this:-
Code: if (!jack_port_type_get_buffer_size) { warning << _("This version of JACK is old - you should upgrade to a newer version") << endmsg; } else { some_var = jack_port_type_get_buffer_size(); }
We link to the latest version of jack, where that symbol is declared like so:-
Code: size_t jack_port_type_get_buffer_size();
One problem is that it doesn't even seem to be an exported symbol (although that wouldn't affect the other platform builds). But apart from that, our customer might have an old copy of jack installed or even no copy! We seem to be making the assumption that if our customer has an up-to-date version, jack_port_type_get_buffer_size will be set to a valid address - but in all other case it'll be magically set to zero (there's nothing in our code that sets it to zero).
I have a specific byte (that is unsigned char) array, that I need to find in a very big file (2GB or so), currently I'm using:
size_t fsFind(FILE* device, byte* data, size_t size) { int SIZE = (1024 > size) ? 1024 : size; byte buffer[SIZE]; int pos = 0; int loc = ftell(device);
[Code] ....
Which seems to find proper result on first use, but on subsequent searches it seems to find invalid positions, is there something I'm doing wrong, or is there a better way?
I am working on a program which creates a large pointer array of numbers and then performs several iterations of operations in them.
Code: int * u = new int[N]; double * nu = new double[N]; int * nud = new int[N]; for (int i=0;i<M;i++){ for (int i=0;i<N;i++){ u[i]=0; nu[i]=0;
[Code]...
If M is small enough then there are no problems in the program. However once M is large enough I get the "unhanded exception":
std::bad_alloc at memory location 0x0026f728..
Since I am just reusing the same arrays, and since I am able to make it through a few iterations, I didn't think it could be a memory issue. If it is, is there a way I can clear the data completely after each iteration?
In C++, the largest int value is 2147483647. So, an integer larger than this cannot be stored and processed as an integer. Similarly, if the sum or product of two positive integers is greater than 2147483647, the result will be incorrect. One way to store and manipulate large integers is to store each individual digit of the number in an array. Write a program that inputs two positive integers of, at most, 20 digits and outputs the sum of the numbers. If the sum of the numbers has more than 20 digits, output the sum with an appropriate message. Your program must, at least, contain a function to read and store a number into an array and another function to output the sum of the numbers. (Hint: Read numbers as strings and store the digits of the number in the reverse order.)
My code handles smaller numbers well enough, but I need the program to be able at least factor 100!.
#include <stdio.h> void factorialOutput(unsigned int &n, int fac[]); unsigned long long factorial(int n); int main(int argc, const char * argv[]) { unsigned int t = 0; int n[101];
I am trying to run a simulation with a large number of objects (mainly arrays and vectors). I am not sure where shall I define my objects: inside or outside of the main() function, like the following two structures:
I know there is a question about scope. But besides this question (which seems have no difference between these two structures here), is there any difference in terms of execution performance or security issue?
I need to parse the file in such a way that I can create a filesystem hierarchy as if I were enumerating files/directories. Ultimately I want to add these to a tree gui control with everything under its proper node without duplicating anything. It should look roughly like so:
dir -file -dir -file -dir -file
I can open the file and add nodes/children to the tree control but how should I go about doing the actual parsing? How can I find a filename and say "this belongs under this node"? I want to do this efficient as possible even if I must use multiple threads.
So I'm attempting to write a program that will parse through a large file (genome sequences) and I'm basically wondering what options I should consider if I wanted to either:
a) store the entire genome in memory and then parse through it b) parse through a file in small portions
If I go with "a", should I just read a file into a vector and then parse through it? And if I go with "b" would I just use an input/output stream?
I've written a Hexadecimal/ASCII chart, and would like to be able to show a larger version of the selected ASCII on screen. Is there a way to read the individual lines of bytes that make up a letter/symbol/number to show on screen.
|0|0|0|0|0|0|0|0| = 0 |0|1|0|0|0|0|1|0| = 66 O O ( These 'O' represent ASCII 219 ) |0|0|1|0|0|1|0|0| = 36 O O |0|0|0|1|1|0|0|0| = 24 OO |0|0|0|1|1|0|0|0| = 24 OO |0|0|1|0|0|1|0|0| = 36 O O |0|1|0|0|0|0|1|0| = 66 O O |0|0|0|0|0|0|0|0| = 0
To make a large graphic X.
I'm using Microsoft 2012 Express, in the console. I don't feel comfortable yet programming in Windows mode.
I'm having problems with progress bar when using a big number in set range. For numbers below 50000 it works very well but for big numbers like 100.000 it doesn't work, it makes 2-3 rounds of animation
Code: int number= 50000; // 50k works well but if i put 100k it won't work (it will animate 2-3 rounds instead of complete one) progressbar1.SetRange(0, number); progressbar1.SetStep(1); for (int i = 0; i < number; ++i) { listcontrol1.InsertItem(i, _T("whatever")); progressbar1.StepIt(); }
Organizing and managing a large database of words building semantic relationships between them.
Questions:
1. Is C++ a good language to manage large databases 2. Is there a better language/software solution that can easily manage my database but return output that I could use in C++ 3. Is C++ a good language for securing and encryption? If not, where should I look?
I have a problem to read a large number of binary files, process them and store them under a new name. The program and routines go very well for 505 files. After reading 506 files, the program now refuses to read the next file. I have 16 Gb of memory and tried to close all other programs and restart the PC. it always stops after 506 files (512 files would be more understanding in a way...).
Here is my code. I have tried many things without success. This is only part of the loop that stops. The if test if (myfile.is_open() returns false by some reason. I can start the process again starting with the file that does not open and then it stops again after 506 files.
char * tfiBlock; ifstream myfile (OrigFilename, ios::in|ios::binary|ios::ate); if (myfile.is_open()) { int lengde = myfile.tellg(); tfiBlock = new char [lengde]; //static char memblock [size];
[Code] .....
Clean up procedure: delete[] tfiBlock;
Are there any limits to how many files that can be opened, or is it maybe someting to be set in the compiler?