diff --git a/src/bytes/bytes_test.go b/src/bytes/bytes_test.go
index 0f6cf49..03f0158 100644
--- a/src/bytes/bytes_test.go
+++ b/src/bytes/bytes_test.go
@@ -693,14 +693,14 @@
for _, r16 := range rt.R16 {
for r := rune(r16.Lo); r <= rune(r16.Hi); r += rune(r16.Stride) {
if r != needle {
- rs = append(rs, rune(r))
+ rs = append(rs, r)
}
}
}
for _, r32 := range rt.R32 {
for r := rune(r32.Lo); r <= rune(r32.Hi); r += rune(r32.Stride) {
if r != needle {
- rs = append(rs, rune(r))
+ rs = append(rs, r)
}
}
}
diff --git a/src/debug/dwarf/entry.go b/src/debug/dwarf/entry.go
index 07b9259..8741479 100644
--- a/src/debug/dwarf/entry.go
+++ b/src/debug/dwarf/entry.go
@@ -554,7 +554,7 @@
case formData16:
val = b.bytes(16)
case formSdata:
- val = int64(b.int())
+ val = b.int()
case formUdata:
val = int64(b.uint())
case formImplicitConst:
diff --git a/src/net/cgo_unix_syscall.go b/src/net/cgo_unix_syscall.go
index 9cfc578..12f15c7 100644
--- a/src/net/cgo_unix_syscall.go
+++ b/src/net/cgo_unix_syscall.go
@@ -85,7 +85,7 @@
func cgoNameinfoPTR(b []byte, sa *syscall.RawSockaddr, salen int) (int, error) {
gerrno, err := unix.Getnameinfo(sa, salen, &b[0], len(b), nil, 0, unix.NI_NAMEREQD)
- return int(gerrno), err
+ return gerrno, err
}
func cgoSockaddrInet4(ip IP) *syscall.RawSockaddr {
diff --git a/src/net/tcpconn_keepalive_test.go b/src/net/tcpconn_keepalive_test.go
index 53d0be0..4bf2f9e 100644
--- a/src/net/tcpconn_keepalive_test.go
+++ b/src/net/tcpconn_keepalive_test.go
@@ -22,7 +22,7 @@
oldCfg KeepAliveConfig
)
testPreHookSetKeepAlive = func(nfd *netFD) {
- oldCfg, errHook = getCurrentKeepAliveSettings(fdType(nfd.pfd.Sysfd))
+ oldCfg, errHook = getCurrentKeepAliveSettings(nfd.pfd.Sysfd)
}
handler := func(ls *localServer, ln Listener) {
@@ -80,7 +80,7 @@
oldCfg KeepAliveConfig
)
testPreHookSetKeepAlive = func(nfd *netFD) {
- oldCfg, errHook = getCurrentKeepAliveSettings(fdType(nfd.pfd.Sysfd))
+ oldCfg, errHook = getCurrentKeepAliveSettings(nfd.pfd.Sysfd)
}
ch := make(chan Conn, 1)
diff --git a/src/runtime/heapdump.go b/src/runtime/heapdump.go
index 5476035..72878d0 100644
--- a/src/runtime/heapdump.go
+++ b/src/runtime/heapdump.go
@@ -460,7 +460,7 @@
continue
}
spf := (*specialfinalizer)(unsafe.Pointer(sp))
- p := unsafe.Pointer(s.base() + uintptr(spf.special.offset))
+ p := unsafe.Pointer(s.base() + spf.special.offset)
dumpfinalizer(p, spf.fn, spf.fint, spf.ot)
}
}
@@ -659,7 +659,7 @@
continue
}
spp := (*specialprofile)(unsafe.Pointer(sp))
- p := s.base() + uintptr(spp.special.offset)
+ p := s.base() + spp.special.offset
dumpint(tagAllocSample)
dumpint(uint64(p))
dumpint(uint64(uintptr(unsafe.Pointer(spp.b))))
diff --git a/src/runtime/mcleanup.go b/src/runtime/mcleanup.go
index c368730..383217a 100644
--- a/src/runtime/mcleanup.go
+++ b/src/runtime/mcleanup.go
@@ -173,14 +173,14 @@
// Reached the end of the linked list. Stop searching at this point.
break
}
- if offset == uintptr(s.offset) && _KindSpecialCleanup == s.kind &&
+ if offset == s.offset && _KindSpecialCleanup == s.kind &&
(*specialCleanup)(unsafe.Pointer(s)).id == c.id {
// The special is a cleanup and contains a matching cleanup id.
*iter = s.next
found = s
break
}
- if offset < uintptr(s.offset) || (offset == uintptr(s.offset) && _KindSpecialCleanup < s.kind) {
+ if offset < s.offset || (offset == s.offset && _KindSpecialCleanup < s.kind) {
// The special is outside the region specified for that kind of
// special. The specials are sorted by kind.
break
diff --git a/src/runtime/mgcmark.go b/src/runtime/mgcmark.go
index a0087ab..8b30604 100644
--- a/src/runtime/mgcmark.go
+++ b/src/runtime/mgcmark.go
@@ -415,7 +415,7 @@
// Don't mark finalized object, but scan it so we retain everything it points to.
// A finalizer can be set for an inner byte of an object, find object beginning.
- p := s.base() + uintptr(spf.special.offset)/s.elemsize*s.elemsize
+ p := s.base() + spf.special.offset/s.elemsize*s.elemsize
// Mark everything that can be reached from
// the object (but *not* the object itself or
diff --git a/src/runtime/mgcsweep.go b/src/runtime/mgcsweep.go
index 1605c21..b72cc46 100644
--- a/src/runtime/mgcsweep.go
+++ b/src/runtime/mgcsweep.go
@@ -553,7 +553,7 @@
siter := newSpecialsIter(s)
for siter.valid() {
// A finalizer can be set for an inner byte of an object, find object beginning.
- objIndex := uintptr(siter.s.offset) / size
+ objIndex := siter.s.offset / size
p := s.base() + objIndex*size
mbits := s.markBitsForIndex(objIndex)
if !mbits.isMarked() {
@@ -561,7 +561,7 @@
// Pass 1: see if it has a finalizer.
hasFinAndRevived := false
endOffset := p - s.base() + size
- for tmp := siter.s; tmp != nil && uintptr(tmp.offset) < endOffset; tmp = tmp.next {
+ for tmp := siter.s; tmp != nil && tmp.offset < endOffset; tmp = tmp.next {
if tmp.kind == _KindSpecialFinalizer {
// Stop freeing of object if it has a finalizer.
mbits.setMarkedNonAtomic()
@@ -573,11 +573,11 @@
// Pass 2: queue all finalizers and clear any weak handles. Weak handles are cleared
// before finalization as specified by the weak package. See the documentation
// for that package for more details.
- for siter.valid() && uintptr(siter.s.offset) < endOffset {
+ for siter.valid() && siter.s.offset < endOffset {
// Find the exact byte for which the special was setup
// (as opposed to object beginning).
special := siter.s
- p := s.base() + uintptr(special.offset)
+ p := s.base() + special.offset
if special.kind == _KindSpecialFinalizer || special.kind == _KindSpecialWeakHandle {
siter.unlinkAndNext()
freeSpecial(special, unsafe.Pointer(p), size)
@@ -589,11 +589,11 @@
}
} else {
// Pass 2: the object is truly dead, free (and handle) all specials.
- for siter.valid() && uintptr(siter.s.offset) < endOffset {
+ for siter.valid() && siter.s.offset < endOffset {
// Find the exact byte for which the special was setup
// (as opposed to object beginning).
special := siter.s
- p := s.base() + uintptr(special.offset)
+ p := s.base() + special.offset
siter.unlinkAndNext()
freeSpecial(special, unsafe.Pointer(p), size)
}
diff --git a/src/runtime/mheap.go b/src/runtime/mheap.go
index d8193dd..cb0d340 100644
--- a/src/runtime/mheap.go
+++ b/src/runtime/mheap.go
@@ -1488,7 +1488,7 @@
s.allocBits = newAllocBits(uintptr(s.nelems))
// Adjust s.limit down to the object-containing part of the span.
- s.limit = s.base() + uintptr(s.elemsize)*uintptr(s.nelems)
+ s.limit = s.base() + s.elemsize*uintptr(s.nelems)
// It's safe to access h.sweepgen without the heap lock because it's
// only ever updated with the world stopped and we run on the
@@ -2152,11 +2152,11 @@
if s == nil {
break
}
- if offset == uintptr(s.offset) && kind == s.kind {
+ if offset == s.offset && kind == s.kind {
found = true
break
}
- if offset < uintptr(s.offset) || (offset == uintptr(s.offset) && kind < s.kind) {
+ if offset < s.offset || (offset == s.offset && kind < s.kind) {
break
}
iter = &s.next
@@ -2323,14 +2323,14 @@
// Reached the end of the linked list. Stop searching at this point.
break
}
- if offset == uintptr(s.offset) && _KindSpecialCheckFinalizer == s.kind &&
+ if offset == s.offset && _KindSpecialCheckFinalizer == s.kind &&
(*specialCheckFinalizer)(unsafe.Pointer(s)).cleanupID == cleanupID {
// The special is a cleanup and contains a matching cleanup id.
*iter = s.next
found = (*specialCheckFinalizer)(unsafe.Pointer(s))
break
}
- if offset < uintptr(s.offset) || (offset == uintptr(s.offset) && _KindSpecialCheckFinalizer < s.kind) {
+ if offset < s.offset || (offset == s.offset && _KindSpecialCheckFinalizer < s.kind) {
// The special is outside the region specified for that kind of
// special. The specials are sorted by kind.
break
@@ -2373,14 +2373,14 @@
// Reached the end of the linked list. Stop searching at this point.
break
}
- if offset == uintptr(s.offset) && _KindSpecialCheckFinalizer == s.kind &&
+ if offset == s.offset && _KindSpecialCheckFinalizer == s.kind &&
(*specialCheckFinalizer)(unsafe.Pointer(s)).cleanupID == cleanupID {
// The special is a cleanup and contains a matching cleanup id.
*iter = s.next
found = s
break
}
- if offset < uintptr(s.offset) || (offset == uintptr(s.offset) && _KindSpecialCheckFinalizer < s.kind) {
+ if offset < s.offset || (offset == s.offset && _KindSpecialCheckFinalizer < s.kind) {
// The special is outside the region specified for that kind of
// special. The specials are sorted by kind.
break
@@ -2476,7 +2476,7 @@
//go:linkname internal_weak_runtime_registerWeakPointer weak.runtime_registerWeakPointer
func internal_weak_runtime_registerWeakPointer(p unsafe.Pointer) unsafe.Pointer {
- return unsafe.Pointer(getOrAddWeakHandle(unsafe.Pointer(p)))
+ return unsafe.Pointer(getOrAddWeakHandle(p))
}
//go:linkname internal_weak_runtime_makeStrongFromWeak weak.runtime_makeStrongFromWeak
diff --git a/src/runtime/slice.go b/src/runtime/slice.go
index 79d3f6c..e31d5dc 100644
--- a/src/runtime/slice.go
+++ b/src/runtime/slice.go
@@ -397,5 +397,5 @@
panicmakeslicelen()
}
cap := roundupsize(uintptr(len), true)
- return unsafe.Slice((*byte)(mallocgc(uintptr(cap), nil, false)), cap)[:len]
+ return unsafe.Slice((*byte)(mallocgc(cap, nil, false)), cap)[:len]
}
diff --git a/src/runtime/traceallocfree.go b/src/runtime/traceallocfree.go
index 70e48ea..b1b6c63 100644
--- a/src/runtime/traceallocfree.go
+++ b/src/runtime/traceallocfree.go
@@ -37,7 +37,7 @@
}
// Emit info.
- w.varint(uint64(trace.minPageHeapAddr))
+ w.varint(trace.minPageHeapAddr)
w.varint(uint64(pageSize))
w.varint(uint64(gc.MinHeapAlign))
w.varint(uint64(fixedStack))
diff --git a/src/runtime/tracebuf.go b/src/runtime/tracebuf.go
index 08a1d46..5adaede 100644
--- a/src/runtime/tracebuf.go
+++ b/src/runtime/tracebuf.go
@@ -183,7 +183,7 @@
// Tolerate a nil mp.
mID := ^uint64(0)
if w.mp != nil {
- mID = uint64(w.mp.procid)
+ mID = w.mp.procid
}
// Write the buffer's header.
@@ -194,7 +194,7 @@
w.byte(byte(w.exp))
}
w.varint(uint64(w.gen))
- w.varint(uint64(mID))
+ w.varint(mID)
w.varint(uint64(ts))
w.traceBuf.lenPos = w.varintReserve()
return w
diff --git a/src/runtime/tracecpu.go b/src/runtime/tracecpu.go
index 092c707..e64ca32 100644
--- a/src/runtime/tracecpu.go
+++ b/src/runtime/tracecpu.go
@@ -258,7 +258,7 @@
if gp != nil {
hdr[1] = gp.goid
}
- hdr[2] = uint64(mp.procid)
+ hdr[2] = mp.procid
// Allow only one writer at a time
for !trace.signalLock.CompareAndSwap(0, 1) {
diff --git a/src/runtime/traceevent.go b/src/runtime/traceevent.go
index 263847b..b0bc4c0 100644
--- a/src/runtime/traceevent.go
+++ b/src/runtime/traceevent.go
@@ -42,7 +42,7 @@
tl.writer().writeProcStatus(uint64(pp.id), procStatus, pp.trace.inSweep).end()
}
if gp := tl.mp.curg; gp != nil && !gp.trace.statusWasTraced(tl.gen) && gp.trace.acquireStatus(tl.gen) {
- tl.writer().writeGoStatus(uint64(gp.goid), int64(tl.mp.procid), goStatus, gp.inMarkAssist, 0 /* no stack */).end()
+ tl.writer().writeGoStatus(gp.goid, int64(tl.mp.procid), goStatus, gp.inMarkAssist, 0 /* no stack */).end()
}
return traceEventWriter{tl}
}
diff --git a/src/runtime/traceruntime.go b/src/runtime/traceruntime.go
index a2775a3..06e36fd 100644
--- a/src/runtime/traceruntime.go
+++ b/src/runtime/traceruntime.go
@@ -457,7 +457,7 @@
// GoStop emits a GoStop event with the provided reason.
func (tl traceLocker) GoStop(reason traceGoStopReason) {
- tl.eventWriter(tracev2.GoRunning, tracev2.ProcRunning).event(tracev2.EvGoStop, traceArg(trace.goStopReasons[tl.gen%2][reason]), tl.stack(0))
+ tl.eventWriter(tracev2.GoRunning, tracev2.ProcRunning).event(tracev2.EvGoStop, trace.goStopReasons[tl.gen%2][reason], tl.stack(0))
}
// GoPark emits a GoBlock event with the provided reason.
@@ -465,7 +465,7 @@
// TODO(mknyszek): Replace traceBlockReason with waitReason. It's silly
// that we have both, and waitReason is way more descriptive.
func (tl traceLocker) GoPark(reason traceBlockReason, skip int) {
- tl.eventWriter(tracev2.GoRunning, tracev2.ProcRunning).event(tracev2.EvGoBlock, traceArg(trace.goBlockReasons[tl.gen%2][reason]), tl.stack(skip))
+ tl.eventWriter(tracev2.GoRunning, tracev2.ProcRunning).event(tracev2.EvGoBlock, trace.goBlockReasons[tl.gen%2][reason], tl.stack(skip))
}
// GoUnpark emits a GoUnblock event.
diff --git a/src/runtime/tracestack.go b/src/runtime/tracestack.go
index 76d6b05..51f3c29 100644
--- a/src/runtime/tracestack.go
+++ b/src/runtime/tracestack.go
@@ -190,7 +190,7 @@
// Emit stack event.
w.byte(byte(tracev2.EvStack))
- w.varint(uint64(node.id))
+ w.varint(node.id)
w.varint(uint64(len(frames)))
for _, frame := range frames {
w.varint(uint64(frame.PC))
diff --git a/src/runtime/tracetype.go b/src/runtime/tracetype.go
index f54f812..613fc88 100644
--- a/src/runtime/tracetype.go
+++ b/src/runtime/tracetype.go
@@ -64,7 +64,7 @@
}
// Emit type.
- w.varint(uint64(node.id))
+ w.varint(node.id)
w.varint(uint64(uintptr(unsafe.Pointer(typ))))
w.varint(uint64(typ.Size()))
w.varint(uint64(typ.PtrBytes))
diff --git a/src/sync/map_test.go b/src/sync/map_test.go
index 05c8135..0d6690c 100644
--- a/src/sync/map_test.go
+++ b/src/sync/map_test.go
@@ -161,7 +161,7 @@
m := new(sync.Map)
for n := int64(1); n <= mapSize; n++ {
- m.Store(n, int64(n))
+ m.Store(n, n)
}
done := make(chan struct{})
diff --git a/src/syscall/dirent.go b/src/syscall/dirent.go
index f6e78d9..c12b119 100644
--- a/src/syscall/dirent.go
+++ b/src/syscall/dirent.go
@@ -33,7 +33,7 @@
case 4:
return uint64(byteorder.BEUint32(b))
case 8:
- return uint64(byteorder.BEUint64(b))
+ return byteorder.BEUint64(b)
default:
panic("syscall: readInt with unsupported size")
}
@@ -48,7 +48,7 @@
case 4:
return uint64(byteorder.LEUint32(b))
case 8:
- return uint64(byteorder.LEUint64(b))
+ return byteorder.LEUint64(b)
default:
panic("syscall: readInt with unsupported size")
}
diff --git a/src/syscall/exec_unix.go b/src/syscall/exec_unix.go
index 4747fa0..3d13d38 100644
--- a/src/syscall/exec_unix.go
+++ b/src/syscall/exec_unix.go
@@ -210,7 +210,7 @@
Close(p[0])
Close(p[1])
releaseForkLock()
- return 0, Errno(err1)
+ return 0, err1
}
releaseForkLock()
@@ -225,7 +225,7 @@
Close(p[0])
if err != nil || n != 0 {
if n == int(unsafe.Sizeof(err1)) {
- err = Errno(err1)
+ err = err1
}
if err == nil {
err = EPIPE
diff --git a/src/syscall/exec_unix_test.go b/src/syscall/exec_unix_test.go
index 3a95356..488b12d 100644
--- a/src/syscall/exec_unix_test.go
+++ b/src/syscall/exec_unix_test.go
@@ -371,7 +371,7 @@
t.Fatalf("could not parse %q as number: %v", s, v)
}
- if v != uint64(orig.Cur) {
+ if v != orig.Cur {
t.Errorf("exec rlimit = %d, want %d", v, orig)
}
}
diff --git a/src/syscall/route_bsd.go b/src/syscall/route_bsd.go
index 46680d6..4e64dcc 100644
--- a/src/syscall/route_bsd.go
+++ b/src/syscall/route_bsd.go
@@ -139,7 +139,7 @@
// - The kernel form appends leading bytes to the prefix field
// to make the <length, prefix> tuple to be conformed with
// the routing message boundary
- l := int(rsaAlignOf(int(b[0])))
+ l := rsaAlignOf(int(b[0]))
if len(b) < l {
return nil, EINVAL
}
diff --git a/src/syscall/syscall.go b/src/syscall/syscall.go
index a46f22d..3657c93 100644
--- a/src/syscall/syscall.go
+++ b/src/syscall/syscall.go
@@ -78,22 +78,22 @@
// Unix returns the time stored in ts as seconds plus nanoseconds.
func (ts *Timespec) Unix() (sec int64, nsec int64) {
- return int64(ts.Sec), int64(ts.Nsec)
+ return ts.Sec, ts.Nsec
}
// Unix returns the time stored in tv as seconds plus nanoseconds.
func (tv *Timeval) Unix() (sec int64, nsec int64) {
- return int64(tv.Sec), int64(tv.Usec) * 1000
+ return tv.Sec, int64(tv.Usec) * 1000
}
// Nano returns the time stored in ts as nanoseconds.
func (ts *Timespec) Nano() int64 {
- return int64(ts.Sec)*1e9 + int64(ts.Nsec)
+ return ts.Sec*1e9 + ts.Nsec
}
// Nano returns the time stored in tv as nanoseconds.
func (tv *Timeval) Nano() int64 {
- return int64(tv.Sec)*1e9 + int64(tv.Usec)*1000
+ return tv.Sec*1e9 + int64(tv.Usec)*1000
}
// Getpagesize and Exit are provided by the runtime.
diff --git a/src/syscall/zsyscall_darwin_arm64.go b/src/syscall/zsyscall_darwin_arm64.go
index 22b0963..d45596e 100644
--- a/src/syscall/zsyscall_darwin_arm64.go
+++ b/src/syscall/zsyscall_darwin_arm64.go
@@ -128,7 +128,7 @@
// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT
func setsockopt(s int, level int, name int, val unsafe.Pointer, vallen uintptr) (err error) {
- _, _, e1 := syscall6(abi.FuncPCABI0(libc_setsockopt_trampoline), uintptr(s), uintptr(level), uintptr(name), uintptr(val), uintptr(vallen), 0)
+ _, _, e1 := syscall6(abi.FuncPCABI0(libc_setsockopt_trampoline), uintptr(s), uintptr(level), uintptr(name), uintptr(val), vallen, 0)
if e1 != 0 {
err = errnoErr(e1)
}
@@ -556,7 +556,7 @@
// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT
func closedir(dir uintptr) (err error) {
- _, _, e1 := syscall(abi.FuncPCABI0(libc_closedir_trampoline), uintptr(dir), 0, 0)
+ _, _, e1 := syscall(abi.FuncPCABI0(libc_closedir_trampoline), dir, 0, 0)
if e1 != 0 {
err = errnoErr(e1)
}
@@ -1257,7 +1257,7 @@
// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT
func readdir_r(dir uintptr, entry *Dirent, result **Dirent) (res Errno) {
- r0, _, _ := syscall(abi.FuncPCABI0(libc_readdir_r_trampoline), uintptr(dir), uintptr(unsafe.Pointer(entry)), uintptr(unsafe.Pointer(result)))
+ r0, _, _ := syscall(abi.FuncPCABI0(libc_readdir_r_trampoline), dir, uintptr(unsafe.Pointer(entry)), uintptr(unsafe.Pointer(result)))
res = Errno(r0)
return
}
@@ -1728,7 +1728,7 @@
_p0 = unsafe.Pointer(&_zero)
}
r0, _, e1 := syscallX(abi.FuncPCABI0(libc_writev_trampoline), uintptr(fd), uintptr(_p0), uintptr(len(iovecs)))
- cnt = uintptr(r0)
+ cnt = r0
if e1 != 0 {
err = errnoErr(e1)
}
@@ -1742,8 +1742,8 @@
// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT
func mmap(addr uintptr, length uintptr, prot int, flag int, fd int, pos int64) (ret uintptr, err error) {
- r0, _, e1 := syscall6X(abi.FuncPCABI0(libc_mmap_trampoline), uintptr(addr), uintptr(length), uintptr(prot), uintptr(flag), uintptr(fd), uintptr(pos))
- ret = uintptr(r0)
+ r0, _, e1 := syscall6X(abi.FuncPCABI0(libc_mmap_trampoline), addr, length, uintptr(prot), uintptr(flag), uintptr(fd), uintptr(pos))
+ ret = r0
if e1 != 0 {
err = errnoErr(e1)
}
@@ -1757,7 +1757,7 @@
// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT
func munmap(addr uintptr, length uintptr) (err error) {
- _, _, e1 := syscall(abi.FuncPCABI0(libc_munmap_trampoline), uintptr(addr), uintptr(length), 0)
+ _, _, e1 := syscall(abi.FuncPCABI0(libc_munmap_trampoline), addr, length, 0)
if e1 != 0 {
err = errnoErr(e1)
}
@@ -1820,7 +1820,7 @@
} else {
_p0 = unsafe.Pointer(&_zero)
}
- _, _, e1 := syscall6(abi.FuncPCABI0(libc_sysctl_trampoline), uintptr(_p0), uintptr(len(mib)), uintptr(unsafe.Pointer(old)), uintptr(unsafe.Pointer(oldlen)), uintptr(unsafe.Pointer(new)), uintptr(newlen))
+ _, _, e1 := syscall6(abi.FuncPCABI0(libc_sysctl_trampoline), uintptr(_p0), uintptr(len(mib)), uintptr(unsafe.Pointer(old)), uintptr(unsafe.Pointer(oldlen)), uintptr(unsafe.Pointer(new)), newlen)
if e1 != 0 {
err = errnoErr(e1)
}
@@ -2016,7 +2016,7 @@
if runtime.GOOS == "ios" {
panic("unimplemented")
}
- _, _, e1 := syscall6(abi.FuncPCABI0(libc_ptrace_trampoline), uintptr(request), uintptr(pid), uintptr(addr), uintptr(data), 0, 0)
+ _, _, e1 := syscall6(abi.FuncPCABI0(libc_ptrace_trampoline), uintptr(request), uintptr(pid), addr, data, 0, 0)
if e1 != 0 {
err = errnoErr(e1)
}
diff --git a/src/time/example_test.go b/src/time/example_test.go
index eeadcdb..05eac86 100644
--- a/src/time/example_test.go
+++ b/src/time/example_test.go
@@ -735,8 +735,8 @@
timeWithoutNanoseconds := time.Date(2000, 2, 1, 12, 13, 14, 0, time.UTC)
withoutNanoseconds := timeWithoutNanoseconds.String()
- fmt.Printf("withNanoseconds = %v\n", string(withNanoseconds))
- fmt.Printf("withoutNanoseconds = %v\n", string(withoutNanoseconds))
+ fmt.Printf("withNanoseconds = %v\n", withNanoseconds)
+ fmt.Printf("withoutNanoseconds = %v\n", withoutNanoseconds)
// Output:
// withNanoseconds = 2000-02-01 12:13:14.000000015 +0000 UTC
diff --git a/src/time/sleep.go b/src/time/sleep.go
index e9cd483..4b7750e 100644
--- a/src/time/sleep.go
+++ b/src/time/sleep.go
@@ -142,7 +142,7 @@
// in Go 1.27 or later.
func NewTimer(d Duration) *Timer {
c := make(chan Time, 1)
- t := (*Timer)(newTimer(when(d), 0, sendTime, c, syncTimer(c)))
+ t := newTimer(when(d), 0, sendTime, c, syncTimer(c))
t.C = c
return t
}
@@ -208,7 +208,7 @@
// be used to cancel the call using its Stop method.
// The returned Timer's C field is not used and will be nil.
func AfterFunc(d Duration, f func()) *Timer {
- return (*Timer)(newTimer(when(d), 0, goFunc, f, nil))
+ return newTimer(when(d), 0, goFunc, f, nil)
}
func goFunc(arg any, seq uintptr, delta int64) {