Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Implemented Coca architecture #2371
base: master
Are you sure you want to change the base?
Implemented Coca architecture #2371
Changes from 8 commits
20cdf41
b8c0ba4
bbe17c4
202526f
367dd39
80ea7d3
3feacb6
f15408f
960873f
33cff54
145d7b5
e8623a9
c9e1ec1
File filter
Filter by extension
Conversations
Jump to
There are no files selected for viewing
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
is the input shape for layer_norm correct?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Fixed it, let me know if it still doesn't look right!
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I think we were doing a conditional import of
keras_nlp
, sokeras_cv
was installable without keras-nlp installed if you are using unrelated features. But that was when the only use was for one tokenizer.keras-cv/keras_cv/models/feature_extractor/clip/clip_model.py
Lines 81 to 85 in 5faae37
We could reconsider if
keras-cv
should hard depend onkeras-nlp
if we want more stuff like this? No strong feelings. @divyashreepathihalli fyiThere was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I think as we support more multi modal models we should depend on KerasNLP. If the tf-text install issue is resolved, we should add it.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Sgtm! Though if we switch to a hard dependency here, we should probably add
keras-nlp
as a dependency insetup.py
(which comes with a transitive dependency ontensorflow-text
andtensorflow
just fyi).There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Do we want to include that in this PR? There's already some imports of Keras-NLP in other places of Keras CV.
If we make it a separate PR, it'll make it easier to rollback if we need to. Considering it's a new dependency, might be worth separating.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This has to be changed to
Example:
since we follow onlyExample
orExamples:
as a standard format.There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
you could keep batch_size as
None
example
self.image_encoder.build((None, self.encoder_width, num_patches))
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Just for my understanding, is there a specific reason to do that over setting the batch_size?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I can't tell if this mask is the right shape or not. Usually you want something
(batch_size, seq_length, seq_length)
(or seq lenth + 1 if that is the effective sequence length. What istext_dim
here?There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
text_dim is the dimensionality of the text embeddings