Reputation: 1307
I'm trying to read a file and I realized that it would fail because I would be trying to read too much data, even though the file is much bigger than what I'm trying to read.
The file is 120 MB, and my ifstream fails at 12967 bytes (even though it starts acting weird at 12801.
Here is code illustrating my problem:
#include <fstream>
#include <iostream>
#include <Windows.h>
using std::ifstream;
using std::cout;
#define CORRECT_SIZE 12800
#define CORRECT_BUT_WIERD 12966
#define INCORRECT_SIZE 12967
bool check_error_bits(ifstream* f);
int main()
{
ifstream myFile("myfile.txt");
char c[CORRECT_SIZE];
char c2[CORRECT_BUT_WIERD];
char c3[INCORRECT_SIZE];
/*
* TEST A (works fine)
*/
myFile.seekg(0, std::ios_base::beg);
myFile.read(c, CORRECT_SIZE);
check_error_bits(&myFile);
cout << myFile.tellg() << std::endl; // Here, tellg() returns 12800
/*
* TEST B (works too, but acts wierd)
*/
myFile.seekg(0, std::ios_base::beg);
myFile.read(c2, CORRECT_BUT_WIERD);
check_error_bits(&myFile);
cout << myFile.tellg() << std::endl; // Here, tellg() returns 16896
/*
* TEST C (FAIL)
*/
myFile.seekg(0, std::ios_base::beg);
myFile.read(c3, INCORRECT_SIZE);
check_error_bits(&myFile);
cout << myFile.tellg() << std::endl; // Here, tellg() returns -1
system("pause");
}
bool check_error_bits(ifstream* f)
{
bool stop = false;
if (f->eof())
{
char msg[500];
strerror_s(msg, errno);
cout << "1: " << msg << std::endl;
}
if (f->fail())
{
char msg[500];
strerror_s(msg, errno);
cout << "2: " << msg << std::endl;
stop = true;
}
if (f->bad())
{
char msg[500];
strerror_s(msg, errno);
cout << "3: " << msg << std::endl;
stop = true;
}
return stop;
}
Trying to read less than 12800 bytes works perfectly well. From 128001 to 12966, it works (although I have not checked if the data is correct), but tellg() returns non-sense. After 12966, read simply fails.
The console output of that program is:
12800
16896
1: No error
2: No error
-1
Press any key to continue . . .
Any help would be appreciated!
Upvotes: 3
Views: 1970
Reputation: 3519
In Windows environment and text files, a character with value 26=0x1A=^Z is taken as "end fo file".
For this reason, if file is not open as binary, an unexpected eof can be received even when file is bigger.
See wikipedia "End of file" (https://en.wikipedia.org/wiki/End-of-file):
In Microsoft's DOS and Windows (and in CP/M and many DEC operating systems), reading from the terminal will never produce an EOF. Instead, programs recognize that the source is a terminal (or other "character device") and interpret a given reserved character or sequence as an end-of-file indicator; most commonly this is an ASCII Control-Z, code 26.
Upvotes: 8
Reputation: 410
You wrote
(although I have not checked if the data is correct)
Do this, because i think the data is correct.
I assume that you're using windows, so: Why can't Explorer decide what size a file is?
Upvotes: 0
Reputation: 182753
You have a 16,896 byte file. The first read works fine. The second read encounters an end of file and can only read 16,896 bytes. Your third read fails because you didn't clear the end of file flag on your stream.
You may also have a 120MB file, but that's not relevant. Check the size of the file in code. (Or it may be 120MB of binary data but you're reading it in text mode.)
Upvotes: 2