Menu

How to Zip and Unzip Files in C#: A Complete Guide

File compression is an essential skill for any C# developer. Whether you're creating backups, reducing storage space, or preparing files for transmission, knowing how to zip and unzip files programmatically can streamline your applications.

This guide walks you through the process using C#'s built-in System.IO.Compression namespace.

Prerequisites

Before getting started, ensure you have:

  • Visual Studio or your preferred C# IDE
  • .NET Framework 4.5 or later
  • Basic understanding of C# file operations

Creating Zip Files in C#

The System.IO.Compression namespace provides the ZipFile and ZipArchive classes for handling zip operations. Here's how to create a zip file:

using System.IO.Compression;

// Create a zip file from a directory
ZipFile.CreateFromDirectory(@"C:\SourceFolder", @"C:\output.zip");

// Create a zip file with custom settings
using (var zipArchive = ZipFile.Open(@"C:\custom.zip", ZipArchiveMode.Create))
{
    zipArchive.CreateEntryFromFile(@"C:\file1.txt", "file1.txt");
    zipArchive.CreateEntryFromFile(@"C:\file2.pdf", "file2.pdf");
}

Extracting Zip Files

Unzipping files is just as straightforward:

// Extract all files to a directory
ZipFile.ExtractToDirectory(@"C:\archive.zip", @"C:\ExtractedFolder");

// Extract specific files
using (var archive = ZipFile.OpenRead(@"C:\archive.zip"))
{
    foreach (var entry in archive.Entries)
    {
        if (entry.Name.EndsWith(".txt"))
        {
            entry.ExtractToFile(Path.Combine(@"C:\ExtractedFolder", entry.Name));
        }
    }
}

Best Practices and Tips

  1. Always use 'using' statements when working with ZipArchive objects to ensure proper resource disposal.
  2. Handle exceptions appropriately, as file operations can fail due to permissions or file access issues.
  3. Check available disk space before extracting large zip files.
  4. Consider using compression levels for optimal file size versus speed trade-offs.

Advanced Features

The System.IO.Compression namespace offers additional features:

// Set compression level
using (var archive = ZipFile.Open(@"C:\compressed.zip", ZipArchiveMode.Create))
{
    archive.CreateEntryFromFile(@"C:\largefile.dat", "largefile.dat", CompressionLevel.Optimal);
}

// Update existing zip files
using (var archive = ZipFile.Open(@"C:\existing.zip", ZipArchiveMode.Update))
{
    archive.CreateEntryFromFile(@"C:\newfile.txt", "newfile.txt");
}

Common Issues and Solutions

  • File Access Errors: Ensure files aren't in use by other processes before zipping/unzipping.
  • Path Too Long: Use shorter file paths or enable long path support in Windows.
  • Out of Memory: Process large files in chunks rather than loading entirely into memory.

Conclusion

Mastering zip operations in C# enables you to create more efficient applications that handle file compression seamlessly. The System.IO.Compression namespace provides all the tools needed for basic to advanced zip operations, making it easy to implement file compression in your C# projects.

Remember to always test your zip operations thoroughly and implement proper error handling to ensure robust file compression functionality in your applications.

3
40

Related

Storing passwords as plain text is dangerous. Instead, you should hash them using a strong, slow hashing algorithm like BCrypt, which includes built-in salting and resistance to brute-force attacks.

Step 1: Install BCrypt NuGet Package

Before using BCrypt, install the BCrypt.Net-Next package:

dotnet add package BCrypt.Net-Next

or via NuGet Package Manager:

Install-Package BCrypt.Net-Next

Step 2: Hash a Password

Use BCrypt.HashPassword() to securely hash a password before storing it:

using BCrypt.Net;

string password = "mySecurePassword123";
string hashedPassword = BCrypt.HashPassword(password);

Console.WriteLine(hashedPassword); // Output: $2a$12$...

Step 3: Verify a Password

To check a user's login attempt, use BCrypt.Verify():

bool isMatch = BCrypt.Verify("mySecurePassword123", hashedPassword);
Console.WriteLine(isMatch); // Output: True

Ensuring proper hashing should be at the top of your list when it comes to building authentication systems.

0
174

Reading a file line by line is useful when handling large files without loading everything into memory at once.

✅ Best Practice: Use File.ReadLines() which is more memory efficient.

Example

foreach (string line in File.ReadLines("file.txt"))
{
    Console.WriteLine(line);
}

Why use ReadLines()?

Reads one line at a time, reducing overall memory usage. Ideal for large files (e.g., logs, CSVs).

Alternative: Use StreamReader (More Control)

For scenarios where you need custom processing while reading the contents of the file:

using (StreamReader reader = new StreamReader("file.txt"))
{
    string? line;
    while ((line = reader.ReadLine()) != null)
    {
        Console.WriteLine(line);
    }
}

Why use StreamReader?

Lets you handle exceptions, encoding, and buffering. Supports custom processing (e.g., search for a keyword while reading).

When to Use ReadAllLines()? If you need all lines at once, use:

string[] lines = File.ReadAllLines("file.txt");

Caution: Loads the entire file into memory—avoid for large files!

2
179

Removing duplicates from a list in C# is a common task, especially when working with large datasets. C# provides multiple ways to achieve this efficiently, leveraging built-in collections and LINQ.

Using HashSet (Fastest for Unique Elements)

A HashSet<T> automatically removes duplicates since it only stores unique values. This is one of the fastest methods:

List<int> numbers = new List<int> { 1, 2, 2, 3, 4, 4, 5 };
numbers = new HashSet<int>(numbers).ToList();
Console.WriteLine(string.Join(", ", numbers)); // Output: 1, 2, 3, 4, 5

Using LINQ Distinct (Concise and Readable)

LINQ’s Distinct() method provides an elegant way to remove duplicates:

List<int> numbers = new List<int> { 1, 2, 2, 3, 4, 4, 5 };
numbers = numbers.Distinct().ToList();
Console.WriteLine(string.Join(", ", numbers)); // Output: 1, 2, 3, 4, 5

Removing Duplicates by Custom Property (For Complex Objects)

When working with objects, DistinctBy() from .NET 6+ simplifies duplicate removal based on a property:

using System.Linq;
using System.Collections.Generic;

class Person
{
    public string Name { get; set; }
    public int Age { get; set; }
}

List<Person> people = new List<Person>
{
    new Person { Name = "Alice", Age = 30 },
    new Person { Name = "Bob", Age = 25 },
    new Person { Name = "Alice", Age = 30 }
};

people = people.DistinctBy(p => p.Name).ToList();
Console.WriteLine(string.Join(", ", people.Select(p => p.Name))); // Output: Alice, Bob

For earlier .NET versions, use GroupBy():

people = people.GroupBy(p => p.Name).Select(g => g.First()).ToList();

Performance Considerations

  • HashSet<T> is the fastest but only works for simple types.
  • Distinct() is easy to use but slower than HashSet<T> for large lists.
  • DistinctBy() (or GroupBy()) is useful for complex objects but may have performance trade-offs.

Conclusion

Choosing the best approach depends on the data type and use case. HashSet<T> is ideal for primitive types, Distinct() is simple and readable, and DistinctBy() (or GroupBy()) is effective for objects.

0
61