we have one file (11 Million) line that is being matched with (10 Billion) line.
the proof of concept we are trying , is to join them on Unix :
All files are delimited and they have composite keys..
could unix be faster than Oracle in This regards..
Please advice
If performance is your concern (rather than data integrity), then a lot more factors are at play than just the platform you are using. For instance the structure of you data might be a factor. For large datasets, databases are almost always better, but if you have a test machine I am keen to find out which of the two methods you prefer and what your observations are.