Unix Timestamp Converter

Convert between Unix timestamps and dates — seconds, milliseconds, microseconds, nanoseconds

Current Timestamp

3/26/2026, 12:08:37 PM

Seconds
1774526917
Milliseconds
1774526917136
Microseconds
1774526917136000
Nanoseconds
1774526917136000000

Timestamp → Date

Date → Timestamp

Working with Unix Timestamps and Date Formats

A Unix timestamp is the number of seconds that have elapsed since January 1, 1970 (the Unix epoch). It provides a simple, timezone-independent way to represent a moment in time as a single integer. Timestamps are used in databases, APIs, log files, and JWT tokens. This converter translates between Unix timestamps and human-readable date formats like ISO 8601, and supports both seconds and milliseconds precision for JavaScript and database compatibility.

Frequently Asked Questions

What is the Unix epoch?
The Unix epoch is the reference point for Unix timestamps: January 1, 1970 at 00:00:00 UTC. All Unix timestamps are measured as the number of seconds (or milliseconds) since this moment. Dates before the epoch are represented as negative numbers.
How do timestamps handle time zones?
Unix timestamps are always in UTC — they represent an absolute moment in time regardless of timezone. When you display a timestamp to a user, you convert it to their local timezone. This makes timestamps ideal for storing dates in databases and transmitting them via APIs.
What is the difference between seconds and milliseconds timestamps?
Unix timestamps traditionally use seconds (10 digits, e.g. 1709913600), but JavaScript Date.now() and many modern APIs use milliseconds (13 digits, e.g. 1709913600000). Always check which precision your system expects to avoid off-by-1000x errors.
What is the Year 2038 problem?
Systems storing timestamps as 32-bit signed integers will overflow on January 19, 2038, wrapping around to a date in 1901. Modern systems use 64-bit integers, which extend the range to billions of years. If you work with legacy systems, verify they use 64-bit timestamps.