This commit is contained in:
Chris Lu 2020-04-20 17:48:06 -07:00
parent 5eb83dfbd3
commit cb3985be70
8 changed files with 15 additions and 15 deletions

View File

@ -11,7 +11,7 @@ import (
) )
func TestCreateAndFind(t *testing.T) { func TestCreateAndFind(t *testing.T) {
filer := filer2.NewFiler(nil, nil, "",0, "", "", nil) filer := filer2.NewFiler(nil, nil, "", 0, "", "", nil)
dir, _ := ioutil.TempDir("", "seaweedfs_filer_test") dir, _ := ioutil.TempDir("", "seaweedfs_filer_test")
defer os.RemoveAll(dir) defer os.RemoveAll(dir)
store := &LevelDBStore{} store := &LevelDBStore{}

View File

@ -11,7 +11,7 @@ import (
) )
func TestCreateAndFind(t *testing.T) { func TestCreateAndFind(t *testing.T) {
filer := filer2.NewFiler(nil, nil, "",0, "", "", nil) filer := filer2.NewFiler(nil, nil, "", 0, "", "", nil)
dir, _ := ioutil.TempDir("", "seaweedfs_filer_test") dir, _ := ioutil.TempDir("", "seaweedfs_filer_test")
defer os.RemoveAll(dir) defer os.RemoveAll(dir)
store := &LevelDB2Store{} store := &LevelDB2Store{}
@ -66,7 +66,7 @@ func TestCreateAndFind(t *testing.T) {
} }
func TestEmptyRoot(t *testing.T) { func TestEmptyRoot(t *testing.T) {
filer := filer2.NewFiler(nil, nil, "",0, "", "", nil) filer := filer2.NewFiler(nil, nil, "", 0, "", "", nil)
dir, _ := ioutil.TempDir("", "seaweedfs_filer_test2") dir, _ := ioutil.TempDir("", "seaweedfs_filer_test2")
defer os.RemoveAll(dir) defer os.RemoveAll(dir)
store := &LevelDB2Store{} store := &LevelDB2Store{}

View File

@ -38,7 +38,7 @@ func (store *MongodbStore) Initialize(configuration util.Configuration, prefix s
store.database = configuration.GetString(prefix + "database") store.database = configuration.GetString(prefix + "database")
store.collectionName = "filemeta" store.collectionName = "filemeta"
poolSize := configuration.GetInt(prefix + "option_pool_size") poolSize := configuration.GetInt(prefix + "option_pool_size")
return store.connection(configuration.GetString(prefix + "uri"), uint64(poolSize)) return store.connection(configuration.GetString(prefix+"uri"), uint64(poolSize))
} }
func (store *MongodbStore) connection(uri string, poolSize uint64) (err error) { func (store *MongodbStore) connection(uri string, poolSize uint64) (err error) {
@ -60,7 +60,7 @@ func (store *MongodbStore) connection(uri string, poolSize uint64) (err error) {
return err return err
} }
func (store *MongodbStore) createIndex (c *mongo.Collection, index mongo.IndexModel, opts *options.CreateIndexesOptions) error { func (store *MongodbStore) createIndex(c *mongo.Collection, index mongo.IndexModel, opts *options.CreateIndexesOptions) error {
_, err := c.Indexes().CreateOne(context.Background(), index, opts) _, err := c.Indexes().CreateOne(context.Background(), index, opts)
return err return err
} }
@ -169,14 +169,14 @@ func (store *MongodbStore) DeleteFolderChildren(ctx context.Context, fullpath ut
func (store *MongodbStore) ListDirectoryEntries(ctx context.Context, fullpath util.FullPath, startFileName string, inclusive bool, limit int) (entries []*filer2.Entry, err error) { func (store *MongodbStore) ListDirectoryEntries(ctx context.Context, fullpath util.FullPath, startFileName string, inclusive bool, limit int) (entries []*filer2.Entry, err error) {
var where = bson.M{"directory": string(fullpath), "name": bson.M{"$gt": startFileName,}} var where = bson.M{"directory": string(fullpath), "name": bson.M{"$gt": startFileName}}
if inclusive { if inclusive {
where["name"] = bson.M{ where["name"] = bson.M{
"$gte": startFileName, "$gte": startFileName,
} }
} }
optLimit := int64(limit) optLimit := int64(limit)
opts := &options.FindOptions{Limit: &optLimit, Sort: bson.M{ "name": 1 }} opts := &options.FindOptions{Limit: &optLimit, Sort: bson.M{"name": 1}}
cur, err := store.connect.Database(store.database).Collection(store.collectionName).Find(ctx, where, opts) cur, err := store.connect.Database(store.database).Collection(store.collectionName).Find(ctx, where, opts)
for cur.Next(ctx) { for cur.Next(ctx) {
var data Model var data Model

View File

@ -18,7 +18,7 @@ type TopicPartition struct {
} }
type TopicLock struct { type TopicLock struct {
sync.Mutex sync.Mutex
cond *sync.Cond cond *sync.Cond
subscriberCount int subscriberCount int
publisherCount int publisherCount int
logBuffer *log_buffer.LogBuffer logBuffer *log_buffer.LogBuffer

View File

@ -49,9 +49,9 @@ func (mc *MessagingClient) NewPublisher(namespace, topic string) (*Publisher, er
doneChan <- err doneChan <- err
return return
} }
if in.Redirect != nil{ if in.Redirect != nil {
} }
if in.Config != nil{ if in.Config != nil {
} }
} }
}() }()

View File

@ -21,11 +21,11 @@ import (
_ "github.com/chrislusf/seaweedfs/weed/filer2/etcd" _ "github.com/chrislusf/seaweedfs/weed/filer2/etcd"
_ "github.com/chrislusf/seaweedfs/weed/filer2/leveldb" _ "github.com/chrislusf/seaweedfs/weed/filer2/leveldb"
_ "github.com/chrislusf/seaweedfs/weed/filer2/leveldb2" _ "github.com/chrislusf/seaweedfs/weed/filer2/leveldb2"
_ "github.com/chrislusf/seaweedfs/weed/filer2/mongodb"
_ "github.com/chrislusf/seaweedfs/weed/filer2/mysql" _ "github.com/chrislusf/seaweedfs/weed/filer2/mysql"
_ "github.com/chrislusf/seaweedfs/weed/filer2/postgres" _ "github.com/chrislusf/seaweedfs/weed/filer2/postgres"
_ "github.com/chrislusf/seaweedfs/weed/filer2/redis" _ "github.com/chrislusf/seaweedfs/weed/filer2/redis"
_ "github.com/chrislusf/seaweedfs/weed/filer2/redis2" _ "github.com/chrislusf/seaweedfs/weed/filer2/redis2"
_ "github.com/chrislusf/seaweedfs/weed/filer2/mongodb"
"github.com/chrislusf/seaweedfs/weed/glog" "github.com/chrislusf/seaweedfs/weed/glog"
"github.com/chrislusf/seaweedfs/weed/notification" "github.com/chrislusf/seaweedfs/weed/notification"
_ "github.com/chrislusf/seaweedfs/weed/notification/aws_sqs" _ "github.com/chrislusf/seaweedfs/weed/notification/aws_sqs"

View File

@ -14,7 +14,7 @@ type OnDiskCacheLayer struct {
diskCaches []*ChunkCacheVolume diskCaches []*ChunkCacheVolume
} }
func NewOnDiskCacheLayer(dir, namePrefix string, diskSizeMB int64, segmentCount int) *OnDiskCacheLayer{ func NewOnDiskCacheLayer(dir, namePrefix string, diskSizeMB int64, segmentCount int) *OnDiskCacheLayer {
volumeCount, volumeSize := int(diskSizeMB/30000), int64(30000) volumeCount, volumeSize := int(diskSizeMB/30000), int64(30000)
if volumeCount < segmentCount { if volumeCount < segmentCount {
@ -58,7 +58,7 @@ func (c *OnDiskCacheLayer) setChunk(needleId types.NeedleId, data []byte) {
} }
func (c *OnDiskCacheLayer) getChunk(needleId types.NeedleId) (data []byte){ func (c *OnDiskCacheLayer) getChunk(needleId types.NeedleId) (data []byte) {
var err error var err error
@ -80,7 +80,7 @@ func (c *OnDiskCacheLayer) getChunk(needleId types.NeedleId) (data []byte){
} }
func (c *OnDiskCacheLayer) shutdown(){ func (c *OnDiskCacheLayer) shutdown() {
for _, diskCache := range c.diskCaches { for _, diskCache := range c.diskCaches {
diskCache.Shutdown() diskCache.Shutdown()

View File

@ -72,4 +72,4 @@ func (logBuffer *LogBuffer) LoopProcessLogData(
// fmt.Printf("sent message ts[%d,%d] size %d\n", startReadTime.UnixNano(), lastReadTime.UnixNano(), batchSize) // fmt.Printf("sent message ts[%d,%d] size %d\n", startReadTime.UnixNano(), lastReadTime.UnixNano(), batchSize)
} }
} }